-

 

 

 




Optimization Online





 

Regularized Nonlinear Acceleration

Damien Scieur(damien.scieur***at***inria.fr)
Francis Bach(francis.bach***at***inria.fr)
Alexandre d'Aspremont(aspremon***at***ens.fr)

Abstract: We describe a convergence acceleration technique for generic optimization problems. Our scheme computes estimates of the optimum from a nonlinear average of the iterates produced by any optimization method. The weights in this average are computed via a simple linear system, whose solution can be updated online. This acceleration scheme runs in parallel to the base algorithm, providing improved estimates of the solution on the fly, while the original optimization method is running. Numerical experiments are detailed on classical classification problems.

Keywords:

Category 1: Convex and Nonsmooth Optimization (Convex Optimization )

Citation: NIPS 2016

Download: [PDF]

Entry Submitted: 05/30/2017
Entry Accepted: 05/30/2017
Entry Last Modified: 05/30/2017

Modify/Update this entry


  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository

 

Submit
Update
Policies
Coordinator's Board
Classification Scheme
Credits
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society