Optimization Online


A Distributionally-robust Approach for Finding Support Vector Machines

Changhyeok Lee(changhyeoklee2014***at***u.northwestern.edu)
Sanjay Mehrotra(mehrotra***at***northwestern.edu)

Abstract: The classical SVM is an optimization problem minimizing the hinge losses of mis-classified samples with the regularization term. When the sample size is small or data has noise, it is possible that the classifier obtained with training data may not generalize well to pop- ulation, since the samples may not accurately represent the true population distribution. We propose a distributionally-robust framework for Support Vector Machines (DR-SVMs). We build an ambiguity set for the population distribution based on samples using the Kan- torovich metric. DR-SVMs search the classifier that minimizes the sum of regularization term and the hinge loss function for the worst-case population distribution among the ambi- guity set. We provide semi-infinite programming formulation of the DR-SVMs and propose a cutting-plane algorithm to solve the problem. Computational results on simulated data and real data from University of California, Irvine Machine Learning Repository show that the DR-SVMs outperform the SVMs in terms of the Area Under Curve (AUC) measures on several test problems.

Keywords: Support Vector Machines (SVMs), Classification, Robust, Distributionally- robust, Kantrorovich metric

Category 1: Applications -- Science and Engineering (Data-Mining )

Category 2: Robust Optimization

Category 3: Stochastic Programming


Download: [PDF]

Entry Submitted: 06/17/2015
Entry Accepted: 06/17/2015
Entry Last Modified: 06/17/2015

Modify/Update this entry

  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository


Coordinator's Board
Classification Scheme
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society