On the sufficient descent property of the Shanno’s conjugate gradient method
详细信息    查看全文
  • 作者:Saman Babaie-Kafaki (1) (2)
  • 关键词:Conjugate gradient algorithm ; Sufficient descent condition ; Line search
  • 刊名:Optimization Letters
  • 出版年:2013
  • 出版时间:April 2013
  • 年:2013
  • 卷:7
  • 期:4
  • 页码:831-837
  • 全文大小:136KB
  • 参考文献:1. Andrei N.: Numerical comparison of conjugate gradient algorithms for unconstrained optimization. Stud. Inform. Control 16(4), 333-52 (2007)
    2. Andrei N.: A scaled BFGS preconditioned conjugate gradient algorithm for unconstrained optimization. Appl. Math. Lett. 20(6), 645-50 (2007) CrossRef
    3. Andrei N.: Scaled conjugate gradient algorithms for unconstrained optimization. Comput. Optim. Appl. 38(3), 401-16 (2007) CrossRef
    4. Andrei N.: Scaled memoryless BFGS preconditioned conjugate gradient algorithm for unconstrained optimization. Optim. Methods Softw. 22(4), 561-71 (2007) CrossRef
    5. Andrei N.: Accelerated scaled memoryless BFGS preconditioned conjugate gradient algorithm for unconstrained optimization. Eur. J. Oper. Res. 204(3), 410-20 (2010) CrossRef
    6. Birgin E., Martínez J.M.: A spectral conjugate gradient method for unconstrained optimization. Appl. Math. Optim. 43(2), 117-28 (2001) CrossRef
    7. Broyden C.G.: The convergence of a class of double-rank minimization algorithms. II. The new algorithm. J. Inst. Math. Appl. 6(1), 222-31 (1970) CrossRef
    8. Dai Y.H., Han J.Y., Liu G.H., Sun D.F., Yin H.X., Yuan Y.X.: Convergence properties of nonlinear conjugate gradient methods. SIAM J. Optim. 10(2), 348-58 (1999)
    9. Dai Y.H., Liao L.Z.: New conjugacy conditions and related nonlinear conjugate gradient methods. Appl. Math. Optim. 43(1), 87-01 (2001) CrossRef
    10. Dai Y.H., Ni Q.: Testing different conjugate gradient methods for large-scale unconstrained optimization. J. Comput. Math. 22(3), 311-20 (2003)
    11. Dai Y.H., Yuan Y.X.: A nonlinear conjugate gradient method with a strong global convergence property. SIAM J. Optim. 10(1), 177-82 (1999) CrossRef
    12. Du D., Pardalos P.M., Wu W.: Mathematical Theory of Optimization. Kluwer Academic Publishers, Dordrecht (2001) CrossRef
    13. Fletcher R.: A new approach to variable metric algorithms. Comput. J. 13(3), 317-22 (1970) CrossRef
    14. Gilbert J.C., Nocedal J.: Global convergence properties of conjugate gradient methods for optimization. SIAM J. Optim. 2(1), 21-2 (1992) CrossRef
    15. Goldfarb D.: A family of variable-metric methods derived by variational means. Math. Comput. 24(109), 23-6 (1970) CrossRef
    16. Hager W.W., Zhang H.: A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM J. Optim. 16(1), 170-92 (2005) CrossRef
    17. Hager W.W., Zhang H.: A survey of nonlinear conjugate gradient methods. Pacific J. Optim. 2(1), 35-8 (2006)
    18. Hestenes M.R., Stiefel E.: Methods of conjugate gradients for solving linear systems. J. Res. Nat. Bureau Standards 49(6), 409-36 (1952) CrossRef
    19. Perry A.: A modified conjugate gradient algorithm. Oper. Res. 26(6), 1073-078 (1976) CrossRef
    20. Powell, M.J.D.: Nonconvex minimization calculations and the conjugate gradient method. In: Griffiths, D.F. (ed.) Numerical Analysis (Dundee, 1983). Lecture Notes in Mathematics, vol. 1066, pp. 122-41. Springer, Berlin (1984)
    21. Shanno D.F.: Conditioning of quasi-Newton methods for function minimization. Math. Comput. 24(111), 647-56 (1970) CrossRef
    22. Shanno D.F.: Conjugate gradient methods with inexact searches. Math. Oper. Res. 3(3), 244-56 (1978) CrossRef
    23. Sun W., Yuan Y.X.: Optimization Theory and Methods: Nonlinear Programming. Springer, (2006)
    24. Wolfe P.: Convergence conditions for ascent methods. SIAM Rev. 11(2), 226-35 (1969) CrossRef
    25. Wolfe P.: Convergence conditions for ascent methods, II. Some corrections. SIAM Rev. 13(2), 185-88 (1971) CrossRef
  • 作者单位:Saman Babaie-Kafaki (1) (2)

    1. Department of Mathematics, Faculty of Mathematics, Statistics and Computer Sciences, Semnan University, P.O. Box 35195-363, Semnan, Iran
    2. School of Mathematics, Institute for Research in Fundamental Sciences (IPM), P.O. Box 19395-5746, Tehran, Iran
  • ISSN:1862-4480
文摘
Satisfying in the sufficient descent condition is a strength of a conjugate gradient method. Here, it is shown that under the Wolfe line search conditions the search directions generated by the memoryless BFGS conjugate gradient algorithm proposed by Shanno satisfy the sufficient descent condition for uniformly convex functions.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700