Optimization Online


A weighted Mirror Descent algorithm for nonsmooth convex optimization problem

Duy V.N. Luong (vu.luong***at***imperial.ac.uk)
Panos Parpas (p.parpas***at***imperial.ac.uk)
Daniel Rueckert (d.rueckert***at***imperial.ac.uk)
Berc Rustem (b.rustem***at***imperial.ac.uk)

Abstract: Large scale nonsmooth convex optimization is a common problem for a range of computational areas including machine learning and computer vision. Problems in these areas contain special domain structures and characteristics. Special treatment of such problem domains, exploiting their structures, can significantly improve the the computational burden. We present a weighted Mirror Descent method to solve optimization problems over a Cartesian product of convex sets. The algorithm uses a nonlinear weighted projection scheme to generate feasible search directions. This projection scheme uses weighting parameters that, eventually, lead to the optimal step-size strategy for every projection on a corresponding subset. We demonstrate the efficiency of the algorithm by solving the Markov Random Fields optimization problem. In particular, we employ the log-entropy distance and the Euclidean distance in the proposed algorithm. Promising experimental results demonstrate the effectiveness of the proposed method.

Keywords: Subgradient Projection, Weighted Projection, Mirror Descent, Markov Random Fields

Category 1: Convex and Nonsmooth Optimization (Nonsmooth Optimization )

Category 2: Convex and Nonsmooth Optimization (Convex Optimization )


Download: [PDF]

Entry Submitted: 03/23/2015
Entry Accepted: 03/23/2015
Entry Last Modified: 09/13/2015

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society