脖子为什么会痒| 膀胱壁增厚是什么原因| 肚子下面是什么部位| loho是什么牌子| 闷骚男是什么意思| 盐酸安罗替尼胶囊主要治疗什么| 减肥吃什么药效果最好| 七月十日是什么日子| 梦见很多小孩是什么意思| 五路财神都叫什么名字| 拉稀吃什么食物好| 什么是玛瑙| 脚掉皮是什么原因| 螃蟹不能和什么一起吃| 牙冠是什么| 多吃菠萝有什么好处| 孕妇d2聚体高是什么原因| 窜稀是什么意思| 皮肤长小肉粒是什么原因| 真棒是什么意思| 吃生南瓜子有什么好处| 不加一笔是什么字| yy什么意思| 爻卦是什么意思| 中暑吃什么食物好| 做梦代表什么生肖| 天气一热身上就痒是什么原因| 妊娠状态是什么意思| 什么是原发性高血压和继发性高血压| 清补凉是什么| 吴亦凡演过什么电影| 两个月没有来月经了是什么原因| 往届毕业生是什么意思| 有腿毛的男人说明什么| 猫叫是什么原因| 夏至吃什么| 肺炎挂什么科| 芒果什么品种最好吃| 造影是什么意思| 氨曲南是什么药| bottle什么意思| 包皮炎用什么药最有效| 阴阳失调是什么意思| 什么三什么四| pf什么意思| 免疫组织化学染色诊断是什么| ipa啤酒是指什么| 奶茶里面的珍珠是什么做的| 大小便失禁是什么原因| 1893年属什么生肖| 浅表性胃炎吃什么药好| 泽五行属什么| 全身皮肤瘙痒是什么原因| 低压低什么原因| bv是什么意思| 眼睛经常充血是什么原因引起的| 什么是预防医学| 出汗少是什么原因| 什么人不能吃海带| nbc是什么意思| 为什么总是犯困想睡觉| 谷维素是什么| 无故流鼻血是什么原因| 阴虱病是什么原因引起的| 生理盐水和食用盐水有什么区别| 冰箱里有什么细菌| 天灵盖是什么意思| 拔完罐需要注意什么| 麻腮风疫苗是预防什么| 没晨勃说明什么问题| 睾丸大是什么原因| 拔了牙可以吃什么| 8月1日是什么节日| 肿瘤是什么病| 什么东西解辣| 怀孕10多天有什么症状| 梦见别人笑什么意思| 什么是应届毕业生| 三级医院是什么意思| 带状疱疹什么引起的| 狗狗气喘吃什么药| 手指头发红是什么原因| halloween是什么意思| 失眠吃什么好睡觉| 货值是什么意思| 普瑞巴林胶囊治什么病| 牛黄安宫丸什么时候吃最好| 耄耋读什么| 混剪是什么意思| 三分三是什么药| 双什么意思| 静脉曲张挂号挂什么科| 梦到孩子丢了是什么征兆| foh是什么意思| 人参有什么作用| 胃酸胃胀吃什么药| 为什么太阳穴疼| 胆固醇是什么东西| 宝宝头爱出汗是什么原因| 为什么没有西京| 22是什么生肖| 青睐什么意思| 罗贯中是什么朝代的| 啵是什么意思| 坐月子吃什么水果好| 经常腹痛什么原因| 什么可以补肾壮阳| miles是什么意思| 1955年属羊的是什么命| 四点底和什么有关| 肩胛骨痛是什么原因| 调理脾胃吃什么好| 三和大神是什么意思| 爸爸的舅舅叫什么| 招财猫是什么品种| 什么是穿刺手术| www是什么| 齐人之福什么意思| 吃什么能快速排便| 洁白丸治什么类型胃病| 唯美什么意思| 肚脐眼位置疼是什么原因| 猪肝不能和什么一起吃| 心力衰竭吃什么药最好| 胆固醇和血脂有什么区别| 冬天喝什么汤| 肚脐下方疼是什么原因| 正直是什么意思| 发膜什么牌子效果最好| 韩红是什么民族| 误喝碘伏有什么伤害吗| 活好的女人有什么表现| 梦见开车是什么意思| 为什么一喝阿胶睡眠就好了| 心包积液吃什么药| 梅核气西医叫什么| 嗓子疼喝什么茶最有效| 不可一世是什么意思| 延年益寿的益是什么意思| 大公鸡衣服是什么牌子| 山水不相逢什么意思| 免疫系统由什么组成| 杜鹃花什么时候开| 蛇缠腰是什么病| 狗刨坑是什么征兆| 山本耀司的品牌叫什么| 白色裤子搭什么颜色上衣| cancer是什么意思| 博士在古代是什么意思| 取环需要做什么检查| 毛的部首是什么| cra是什么| 萤火虫为什么会发光简单回答| 2030年属什么生肖| 右乳钙化灶是什么意思| 诺言背叛诺言是什么歌| 血管检查什么方法最好| 血象高会导致什么后果| 痔疮的症状是什么| 安慰是什么意思| 米是什么结构| 慢性胃炎伴胆汁反流是什么意思| 遣返回国有什么后果| cp什么意思| 什么的星空| 遥不可及是什么意思| 寻常是什么意思| 榴莲为什么苦| 白细胞高说明什么问题| 内衣什么品牌最好| 闰六月给父母买什么| 宫颈炎盆腔炎吃什么药效果最好| 头上的旋有什么说法| 足跟疼挂什么科室| 冬天开什么花| 属虎和什么属相相冲| 梦见鳄鱼是什么预兆| 成吉思汗是什么族| 为什么要打胰岛素| 什么什么不舍| pep是什么意思| 牛建读什么| 孕期同房需要注意什么| 什么食物含磷高| 什么东西吃了壮阳| 思密达是什么意思| 菠萝蜜什么时候成熟| 宝宝拉肚子有粘液是什么原因| 乙肝病毒是什么| 三级手术是什么意思| 避孕套和安全套有什么区别| 斑秃吃什么药| 洛阳有什么好玩的| 女人脚浮肿是什么原因| 儿女情长英雄气短是什么意思| 胃饱胀是什么原因| 低密度脂蛋白偏高是什么意思| 拉屎很臭是什么原因| 动脉硬化吃什么可以软化血管| 村里入党需要什么条件| 大龄补贴需要什么条件| 头脑简单是什么生肖| 人肉搜索是什么意思| 心脏早搏是什么原因| 失败是成功之母是什么意思| uin是什么意思| 嗜睡是什么症状| 扁桃体长什么样子| 男人为什么会遗精| 老白茶属于什么茶| 什么是中医| 什么茶女人长期喝最好| 王字加一笔是什么字| 猫有什么特点| 教师节唱什么歌| 平均血小板体积偏低是什么意思| 蓝营绿营什么意思| 葛根有什么作用| 子宫为什么长肌瘤| 为什么会反胃想吐| 前夕是什么意思| 戊是什么意思| 一鸣惊人指什么生肖| 对虾是什么虾| 老板喜欢什么样的员工| 颜控什么意思| 费心是什么意思| 一直耳鸣是什么原因引起的| 霉菌性阴道炎是什么症状| 属蛇的是什么星座| 肌酐升高是什么原因| 什么是洁癖| 鹿几念什么| 贵姓是什么意思| 广西属于什么气候| 切除一侧输卵管对女性有什么影响| swan什么意思| 过敏了吃什么药| 胖子从12楼掉下来会变什么| 尿酸高会出现什么症状| 傻狍子为什么叫傻狍子| 天才是指什么生肖| 烤瓷牙是什么意思| 昵称是什么| 陈宝莲为什么自杀| 肺部炎症用什么药最好| 圣罗兰为什么叫杨树林| h的大写字母是什么| mar什么意思| 乳腺增生应该注意些什么| 什么是佛| 夏天吃什么解暑| 提篮子是什么意思| 细胞角蛋白19片段是什么意思| 肾结水有什么危害| 尿白细胞加减什么意思| 五月份什么星座| 养肝护肝吃什么药效果最好| 胆固醇高有什么症状| 豆浆和什么不能一起吃| 鸡的贵人是什么生肖| 子孙满堂是什么生肖| 爸爸的姐姐叫什么| 谦虚什么意思| 百度Jump to content

全屋定制状况频发 槽点都在这儿

From Wikipedia, the free encyclopedia
百度 国家多个部门近期联合下发的《关于开展放心消费创建活动 营造安全放心消费环境的指导意见》就指出,要进一步提升服务领域质量,引导网络交易、网络文化、数字内容等服务消费领域经营者诚信经营,有效规范服务行业市场秩序。

Limited-memory BFGS (L-BFGS or LM-BFGS) is an optimization algorithm in the collection of quasi-Newton methods that approximates the Broyden–Fletcher–Goldfarb–Shanno algorithm (BFGS) using a limited amount of computer memory.[1] It is a popular algorithm for parameter estimation in machine learning.[2][3] The algorithm's target problem is to minimize over unconstrained values of the real-vector where is a differentiable scalar function.

Like the original BFGS, L-BFGS uses an estimate of the inverse Hessian matrix to steer its search through variable space, but where BFGS stores a dense approximation to the inverse Hessian (n being the number of variables in the problem), L-BFGS stores only a few vectors that represent the approximation implicitly. Due to its resulting linear memory requirement, the L-BFGS method is particularly well suited for optimization problems with many variables. Instead of the inverse Hessian Hk, L-BFGS maintains a history of the past m updates of the position x and gradient ?f(x), where generally the history size m can be small (often ). These updates are used to implicitly do operations requiring the Hk-vector product.

Algorithm

[edit]

The algorithm starts with an initial estimate of the optimal value, , and proceeds iteratively to refine that estimate with a sequence of better estimates . The derivatives of the function are used as a key driver of the algorithm to identify the direction of steepest descent, and also to form an estimate of the Hessian matrix (second derivative) of .

L-BFGS shares many features with other quasi-Newton algorithms, but is very different in how the matrix-vector multiplication is carried out, where is the approximate Newton's direction, is the current gradient, and is the inverse of the Hessian matrix. There are multiple published approaches using a history of updates to form this direction vector. Here, we give a common approach, the so-called "two loop recursion."[4][5]

We take as given , the position at the k-th iteration, and where is the function being minimized, and all vectors are column vectors. We also assume that we have stored the last m updates of the form

.

We define , and will be the 'initial' approximate of the inverse Hessian that our estimate at iteration k begins with.

The algorithm is based on the BFGS recursion for the inverse Hessian as

For a fixed k we define a sequence of vectors as and . Then a recursive algorithm for calculating from is to define and . We also define another sequence of vectors as . There is another recursive algorithm for calculating these vectors which is to define and then recursively define and . The value of is then our ascent direction.

Thus we can compute the descent direction as follows:


This formulation gives the search direction for the minimization problem, i.e., . For maximization problems, one should thus take -z instead. Note that the initial approximate inverse Hessian is chosen as a diagonal matrix or even a multiple of the identity matrix since this is numerically efficient.

The scaling of the initial matrix ensures that the search direction is well scaled and therefore the unit step length is accepted in most iterations. A Wolfe line search is used to ensure that the curvature condition is satisfied and the BFGS updating is stable. Note that some software implementations use an Armijo backtracking line search, but cannot guarantee that the curvature condition will be satisfied by the chosen step since a step length greater than may be needed to satisfy this condition. Some implementations address this by skipping the BFGS update when is negative or too close to zero, but this approach is not generally recommended since the updates may be skipped too often to allow the Hessian approximation to capture important curvature information. Some solvers employ so called damped (L)BFGS update which modifies quantities and in order to satisfy the curvature condition.

The two-loop recursion formula is widely used by unconstrained optimizers due to its efficiency in multiplying by the inverse Hessian. However, it does not allow for the explicit formation of either the direct or inverse Hessian and is incompatible with non-box constraints. An alternative approach is the compact representation, which involves a low-rank representation for the direct and/or inverse Hessian.[6] This represents the Hessian as a sum of a diagonal matrix and a low-rank update. Such a representation enables the use of L-BFGS in constrained settings, for example, as part of the SQP method.

Applications

[edit]

L-BFGS has been called "the algorithm of choice" for fitting log-linear (MaxEnt) models and conditional random fields with -regularization.[2][3]

Variants

[edit]

Since BFGS (and hence L-BFGS) is designed to minimize smooth functions without constraints, the L-BFGS algorithm must be modified to handle functions that include non-differentiable components or constraints. A popular class of modifications are called active-set methods, based on the concept of the active set. The idea is that when restricted to a small neighborhood of the current iterate, the function and constraints can be simplified.

L-BFGS-B

[edit]

The L-BFGS-B algorithm extends L-BFGS to handle simple box constraints (aka bound constraints) on variables; that is, constraints of the form lixiui where li and ui are per-variable constant lower and upper bounds, respectively (for each xi, either or both bounds may be omitted).[7][8] The method works by identifying fixed and free variables at every step (using a simple gradient method), and then using the L-BFGS method on the free variables only to get higher accuracy, and then repeating the process.

OWL-QN

[edit]

Orthant-wise limited-memory quasi-Newton (OWL-QN) is an L-BFGS variant for fitting -regularized models, exploiting the inherent sparsity of such models.[3] It minimizes functions of the form

where is a differentiable convex loss function. The method is an active-set type method: at each iterate, it estimates the sign of each component of the variable, and restricts the subsequent step to have the same sign. Once the sign is fixed, the non-differentiable term becomes a smooth linear term which can be handled by L-BFGS. After an L-BFGS step, the method allows some variables to change sign, and repeats the process.

O-LBFGS

[edit]

Schraudolph et al. present an online approximation to both BFGS and L-BFGS.[9] Similar to stochastic gradient descent, this can be used to reduce the computational complexity by evaluating the error function and gradient on a randomly drawn subset of the overall dataset in each iteration. It has been shown that O-LBFGS has a global almost sure convergence [10] while the online approximation of BFGS (O-BFGS) is not necessarily convergent.[11]

Implementation of variants

[edit]

Notable open source implementations include:

  • ALGLIB implements L-BFGS in C++ and C# as well as a separate box/linearly constrained version, BLEIC.
  • R's optim general-purpose optimizer routine uses the L-BFGS-B method.
  • SciPy's optimization module's minimize method also includes an option to use L-BFGS-B.
  • Julia's Optim.jl also implements the L-BFGS and L-BFGS-B algorithm[12].

Notable non open source implementations include:

  • The L-BFGS-B variant also exists as ACM TOMS algorithm 778.[8][13] In February 2011, some of the authors of the original L-BFGS-B code posted a major update (version 3.0).
  • A reference implementation in Fortran 77 (and with a Fortran 90 interface).[14][15] This version, as well as older versions, has been converted to many other languages.
  • An OWL-QN C++ implementation by its designers.[3][16]

Works cited

[edit]
  1. ^ Liu, D. C.; Nocedal, J. (1989). "On the Limited Memory Method for Large Scale Optimization". Mathematical Programming B. 45 (3): 503–528. CiteSeerX 10.1.1.110.6443. doi:10.1007/BF01589116. S2CID 5681609.
  2. ^ a b Malouf, Robert (2002). "A comparison of algorithms for maximum entropy parameter estimation". Proceedings of the Sixth Conference on Natural Language Learning (CoNLL-2002). pp. 49–55. doi:10.3115/1118853.1118871.
  3. ^ a b c d Andrew, Galen; Gao, Jianfeng (2007). "Scalable training of L?-regularized log-linear models". Proceedings of the 24th International Conference on Machine Learning. doi:10.1145/1273496.1273501. ISBN 9781595937933. S2CID 5853259.
  4. ^ Matthies, H.; Strang, G. (1979). "The solution of non linear finite element equations". International Journal for Numerical Methods in Engineering. 14 (11): 1613–1626. Bibcode:1979IJNME..14.1613M. doi:10.1002/nme.1620141104.
  5. ^ Nocedal, J. (1980). "Updating Quasi-Newton Matrices with Limited Storage". Mathematics of Computation. 35 (151): 773–782. doi:10.1090/S0025-5718-1980-0572855-7.
  6. ^ Byrd, R. H.; Nocedal, J.; Schnabel, R. B. (1994). "Representations of Quasi-Newton Matrices and their use in Limited Memory Methods". Mathematical Programming. 63 (4): 129–156. doi:10.1007/BF01582063. S2CID 5581219.
  7. ^ Byrd, R. H.; Lu, P.; Nocedal, J.; Zhu, C. (1995). "A Limited Memory Algorithm for Bound Constrained Optimization". SIAM J. Sci. Comput. 16 (5): 1190–1208. Bibcode:1995SJSC...16.1190B. doi:10.1137/0916069. S2CID 6398414.
  8. ^ a b Zhu, C.; Byrd, Richard H.; Lu, Peihuang; Nocedal, Jorge (1997). "L-BFGS-B: Algorithm 778: L-BFGS-B, FORTRAN routines for large scale bound constrained optimization". ACM Transactions on Mathematical Software. 23 (4): 550–560. doi:10.1145/279232.279236. S2CID 207228122.
  9. ^ Schraudolph, N.; Yu, J.; Günter, S. (2007). A stochastic quasi-Newton method for online convex optimization. AISTATS.
  10. ^ Mokhtari, A.; Ribeiro, A. (2015). "Global convergence of online limited memory BFGS" (PDF). Journal of Machine Learning Research. 16: 3151–3181. arXiv:1409.2045.
  11. ^ Mokhtari, A.; Ribeiro, A. (2014). "RES: Regularized Stochastic BFGS Algorithm". IEEE Transactions on Signal Processing. 62 (23): 6089–6104. arXiv:1401.7625. Bibcode:2014ITSP...62.6089M. CiteSeerX 10.1.1.756.3003. doi:10.1109/TSP.2014.2357775. S2CID 15214938.
  12. ^ "Official Documentation of Optim.jl". Documentation Optim.jl.
  13. ^ "TOMS Home". toms.acm.org.
  14. ^ Morales, J. L.; Nocedal, J. (2011). "Remark on "algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound constrained optimization"". ACM Transactions on Mathematical Software. 38: 1–4. doi:10.1145/2049662.2049669. S2CID 16742561.
  15. ^ "L-BFGS-B Nonlinear Optimization Code". users.iems.northwestern.edu.
  16. ^ "Orthant-Wise Limited-memory Quasi-Newton Optimizer for L1-regularized Objectives". Microsoft Download Center.

Further reading

[edit]
38节送什么礼物 榻榻米是什么 女兔配什么属相最好 barry什么意思 为什么喝纯牛奶会拉肚子
月经过后腰酸疼是什么原因 属鸡的本命佛是什么佛 观音成道日是什么意思 太阳穴凹陷是什么原因 青梅是什么水果
10月21日是什么星座 脑委缩有什么症状 66.66红包代表什么意思 做梦吃面条是什么预兆 驴血是什么颜色
安逸是什么意思 甲流吃什么药 小孩吃什么有营养 半身不遂是什么原因引起的 中蛊什么意思
完美收官什么意思hcv8jop8ns5r.cn 腋下疣是什么原因造成的hcv8jop1ns0r.cn 裸婚什么意思hcv9jop0ns8r.cn 姑娘是什么意思hcv9jop7ns2r.cn 无痛人流后吃什么对身体恢复比较好hcv9jop3ns6r.cn
妇科彩超主要检查什么hcv8jop5ns9r.cn 请人帮忙用什么词clwhiglsz.com 蓝色加红色是什么颜色gysmod.com kissme什么意思hcv9jop5ns6r.cn 什么食物补锌效果最好sanhestory.com
大枣吃多了有什么危害luyiluode.com 飞蚊症是什么引起的hcv7jop6ns8r.cn 地盆是一种什么病hcv8jop8ns4r.cn 客源是什么意思hcv9jop4ns0r.cn 脚手发热是什么原因hcv9jop1ns5r.cn
bigbang是什么意思hcv8jop0ns3r.cn 短阵房速是什么意思hcv7jop7ns2r.cn 来大姨妈吃什么hcv9jop4ns0r.cn 每晚做梦是什么原因hcv9jop4ns8r.cn sle是什么病的缩写hcv8jop6ns7r.cn
百度