-

 

 

 




Optimization Online





 

A Combined Class of Self-Scaling and Modified Quasi-Newton Methods

Mehiddin Al-Baali(alaali***at***squ.edu.om)
Humaid Khalfan(humaidf***at***uaeu.ac.ae)

Abstract: Techniques for obtaining safely positive definite Hessian approximations with self-scaling and modified quasi-Newton updates are combined to obtain `better' curvature approximations in line search methods for unconstrained optimization. It is shown that this class of methods, like the BFGS method has global and superlinear convergence for convex functions. Numerical experiments with this class, using the well-known quasi-Newton BFGS, DFP and a modified SR1 updates, are presented to illustrate advantages of the new techniques. These experiments show that the performance of several combined methods are substantially better than that of the standard BFGS method. These experiments show that the performance of several combined methods are substantially better than that of the standard BFGS method. Similar improvements are also obtained if the simple sufficient function reduction condition on the steplength is used instead of the strong Wolfe conditions.

Keywords: Unconstrained optimization, modified quasi-Newton updates, self-scaling technique, line-search framework.

Category 1: Nonlinear Optimization

Category 2: Nonlinear Optimization (Unconstrained Optimization )

Citation: DOMAS 09/1, Sultan Qaboos University, Oman

Download: [Postscript][PDF]

Entry Submitted: 07/02/2009
Entry Accepted: 07/02/2009
Entry Last Modified: 07/02/2009

Modify/Update this entry


  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository

 

Submit
Update
Policies
Coordinator's Board
Classification Scheme
Credits
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Programming Society