Optimization Online


Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization

Y Narushima(narushima***at***fukushima-nct.ac.jp)
H Yabe(yabe***at***rs.kagu.tus.ac.jp)

Abstract: Conjugate gradient methods have been paid attention to, because they can be directly applied to large-scale unconstrained optimization problems. In order to incorporate second order information of the objective function into conjugate gradient methods, Dai and Liao (2001) proposed a conjugate gradient method based on the secant condition. However, their method does not necessarily generate a descent search direction. On the other hand, Hager and Zhang (2005) proposed another conjugate gradient method which always generates a descent search direction. In this paper, combining Dai-Liao's idea and Hager-Zhang's idea, we propose conjugate gradient methods based on secant conditions that generate descent search directions. In addition, we prove global convergence properties of the proposed methods. Finally, preliminary numerical results are given.

Keywords: Unconstrained optimization, conjugate gradient method, descent search direction, secant condition, global convergence.

Category 1: Nonlinear Optimization (Unconstrained Optimization )

Citation: Department of Mathematical Information Science, Tokyo University of Science. 1-3, Kagurazaka, Shinjuku-ku, Tokyo 162-8601, Japan. September 28, 2011

Download: [PDF]

Entry Submitted: 09/28/2011
Entry Accepted: 09/28/2011
Entry Last Modified: 09/28/2011

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society