Optimization Online


Convergence rate and iteration complexity on the alternating direction method of multipliers with a substitution procedure for separable convex programming

Bingsheng He(hebma***at***nju.edu.cn)
Min Tao(taomin0903***at***gmail.com)
Xiaoming Yuan(xmyuan***at***hkbu.edu.hk)

Abstract: Recently, in [17] we have showed the first possibility of combining the Douglas-Rachford alternating direction method of multipliers (ADMM) with a Gaussian back substitution procedure for solving a convex minimization model with a general separable structure. This paper is a further study on theoretical aspects of this theme. We first derive a general algorithmic framework to combine ADMM with either a forward or backward substitution procedure. Then, we show that convergence of this framework can be easily proved from contraction perspective, and its local linear convergence rate is provable if certain standard error bound condition is assumed. Without such an error bound assumption, we can still estimate the worst-case iteration complexity for this framework in both ergodic and nonergodic senses.

Keywords: Convex programming, Alternating direction method of multipliers, Convergence rate, Iteration complexity, Contraction methods

Category 1: Convex and Nonsmooth Optimization (Convex Optimization )


Download: [PDF]

Entry Submitted: 09/19/2012
Entry Accepted: 09/19/2012
Entry Last Modified: 09/19/2012

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society