A new conjugate gradient method with guaranteed descent and an efficient line search

被引:829
作者
Hager, WW [1 ]
Zhang, HC [1 ]
机构
[1] Univ Florida, Dept Math, Gainesville, FL 32611 USA
关键词
conjugate gradient method; unconstrained optimization; convergence; line search; Wolfe conditions;
D O I
10.1137/030601880
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
A new nonlinear conjugate gradient method and an associated implementation, based on an inexact line search, are proposed and analyzed. With exact line search, our method reduces to a nonlinear version of the Hestenes - Stiefel conjugate gradient scheme. For any ( inexact) line search, our scheme satisfies the descent condition g(k)(T) d(k) <= - 7/8 parallel to g(k)parallel to(2). Moreover, a global convergence result is established when the line search fulfills the Wolfe conditions. A new line search scheme is developed that is efficient and highly accurate. Efficiency is achieved by exploiting properties of linear interpolants in a neighborhood of a local minimizer. High accuracy is achieved by using a convergence criterion, which we call the "approximate Wolfe" conditions, obtained by replacing the sufficient decrease criterion in the Wolfe conditions with an approximation that can be evaluated with greater precision in a neighborhood of a local minimum than the usual sufficient decrease criterion. Numerical comparisons are given with both L-BFGS and conjugate gradient methods using the unconstrained optimization problems in the CUTE library.
引用
收藏
页码:170 / 192
页数:23
相关论文
共 41 条
[41]  
韩继业, 2001, Acta Mathematicae Applicatae Sinica, V17, P38