Optimization Online


On the Information-Adaptive Variants of the ADMM: an Iteration Complexity Perspective

Xiang Gao(gaoxx460***at***umn.edu)
Bo Jiang(isyebojiang***at***163.com)
Shuzhong Zhang(zhangs***at***umn.edu)

Abstract: Designing algorithms for an optimization model often amounts to maintaining a balance between the degree of information to request from the model on the one hand, and the computational speed to expect on the other hand. Naturally, the more information is available, the faster one can expect the algorithm to converge. The popular algorithm of ADMM demands that objective function is easy to optimize once the coupled constraints are shifted to the objective with multipliers. However, in many applications this assumption does not hold; instead, only some noisy estimations of the gradient of the objective -- or even only the objective itself -- are available. This paper aims to bridge this gap. We present a suite of variants of the ADMM, where the trade-offs between the required information on the objective and the computational complexity are explicitly given. The new variants allow the method to be applicable on a much broader class of problems where only noisy estimations of the gradient or the function values are accessible, yet the flexibility is achieved without sacrificing the computational complexity bounds.

Keywords: alternating direction method of multipliers (ADMM), iteration complexity, stochastic approximation, first-order method, direct method.

Category 1: Convex and Nonsmooth Optimization (Convex Optimization )


Download: [PDF]

Entry Submitted: 11/07/2014
Entry Accepted: 11/08/2014
Entry Last Modified: 11/07/2014

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society