Optimization Online


Convergence Analysis of Block Majorize-Minimize Subspace Approaches

Jean-Baptiste Fest (jean-baptiste.fest***at***inria.fr)
Emilie Chouzenoux (emilie.chouzenoux***at***centralesupelec.fr)

Abstract: Majorization-Minimization (MM) consists of a class of efficient and effective optimization algorithms that benefit from solid theoretical foundations. MM methods have shown their great ability to tackle efficiently challenging optimization problems from signal processing, image processing, inverse problems and machine learning. When processing large amount of data/variable, as it may happen in 3D image processing, video processing, or spectroscopy, using standard optimization algorithms becomes difficult due to memory limitation issues. To limit the dependence of an optimization algorithm on the dimension of the problem, block alternating algorithms have been developed where, at each iteration, only a subset of the variables is updated. This paper provides new insights about the convergence guarantees of block alternating MM schemes, in the challenging non-convex setting, and when subspace acceleration is possibly incorporated to improve the convergence rate.

Keywords: block alternating optimization, memory gradient methods, majorization-minimization, non-convex optimization

Category 1: Nonlinear Optimization (Unconstrained Optimization )

Category 2: Applications -- Science and Engineering


Download: [PDF]

Entry Submitted: 12/06/2021
Entry Accepted: 12/06/2021
Entry Last Modified: 02/05/2022

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society