Optimization Online


Distributed Block-diagonal Approximation Methods for Regularized Empirical Risk Minimization

Ching-pei Lee (ching-pei***at***cs.wisc.edu)
Kai-Wei Chang (kw***at***kwchang.net)

Abstract: Designing distributed algorithms for empirical risk minimization (ERM) has become an active research topic in recent years because of the practical need to deal with the huge volume of data. In this paper, we propose a general framework for training an ERM model via solving its dual problem in parallel over multiple machines. Our method provides a versatile approach for many large-scale machine learning problems, including linear binary/multi-class classification, regression, and structured prediction. Comparing with existing approaches, we show that our method has faster convergence under weaker conditions both theoretically and empirically.

Keywords: Distributed optimization, Regularized empirical risk minimization, Linear convergence

Category 1: Convex and Nonsmooth Optimization (Convex Optimization )

Category 2: Convex and Nonsmooth Optimization (Nonsmooth Optimization )


Download: [PDF]

Entry Submitted: 06/04/2017
Entry Accepted: 06/04/2017
Entry Last Modified: 11/16/2019

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society