Optimization Online


Using gradient directions to get global convergence of Newton-type methods

Daniela di Serafino (daniela.diserafino***at***unicampania.it)
Gerardo Toraldo (toraldo***at***unina.it)
Marco Viola (marco.viola***at***unicampania.it)

Abstract: The renewed interest in Steepest Descent (SD) methods following the work of Barzilai and Borwein [IMA Journal of Numerical Analysis, 8 (1988)] has driven us to consider a globalization strategy based on SD, which is applicable to any line-search method. In particular, we combine Newton-type directions with scaled SD steps to have suitable descent directions. Scaling the SD directions with a suitable step length makes a significant difference with respect to similar globalization approaches, in terms of both theoretical features and computational behavior. We apply our strategy to Newton's method and the BFGS method, with computational results that appear interesting compared with the results of well-established globalization strategies devised ad hoc for those methods.

Keywords: Newton-type methods; globalization strategies; steepest descent step

Category 1: Nonlinear Optimization (Unconstrained Optimization )


Download: [PDF]

Entry Submitted: 04/02/2020
Entry Accepted: 04/02/2020
Entry Last Modified: 06/23/2020

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society