Optimization Online


Convergence and Convergence Rate of Stochastic Gradient Search in the Case of Multiple and Non-Isolated Extrema

Vladislav Tadic (v.b.tadic***at***bristol.ac.uk)

Abstract: The asymptotic behavior of stochastic gradient algorithms is studied. Relying on some results of differential geometry (Lojasiewicz gradient inequality), the almost sure point-convergence is demonstrated and relatively tight almost sure bounds on the convergence rate are derived. In sharp contrast to all existing result of this kind, the asymptotic results obtained here do not require the objective function (associated with the stochastic gradient search) to have an isolated minimum at which the Hessian of the objective function is strictly positive definite. Using the obtained results, the asymptotic behavior of recursive prediction error identification methods is analyzed. The convergence and convergence rate of supervised learning algorithms are also studied relying on these results.

Keywords: Stochastic gradient search, point-convergence, convergence rate, Lojasiewicz gradient inequality, system identification, machine learning

Category 1: Stochastic Programming

Category 2: Applications -- Science and Engineering (Control Applications )


Download: [PDF]

Entry Submitted: 07/07/2009
Entry Accepted: 07/07/2009
Entry Last Modified: 07/17/2009

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Programming Society