-

 

 

 




Optimization Online





 

Training Structured Neural Networks Through Manifold Identification and Variance Reduction

Zih-Syuan Huang (efth978jeff***at***gmail.com)
Ching-pei Lee (leechingpei***at***gmail.com)

Abstract: This paper proposes an algorithm, RMDA, for training neural networks (NNs) with a regularization term for promoting desired structures. RMDA does not incur computation additional to proximal SGD with momentum, and achieves variance reduction without requiring the objective function to be of the finite-sum form. Through the tool of manifold identification from nonlinear optimization, we prove that after a finite number of iterations, all iterates of RMDA possess a desired structure identical to that induced by the regularizer at the stationary point of asymptotic convergence, even in the presence of engineering tricks like data augmentation and dropout that complicate the training process. Experiments on training NNs with structured sparsity confirm that variance reduction is necessary for such an identification, and show that RMDA thus significantly outperforms existing methods for this task. For unstructured sparsity, RMDA also outperforms a state-of-the-art pruning method, validating the benefits of training structured NNs through regularization. Implementation of RMDA is available at https://www.github.com/zihsyuan1214/rmda.

Keywords: manifold identification; dual averaging; variance reduction; deep learning; regularized optimization

Category 1: Nonlinear Optimization

Category 2: Convex and Nonsmooth Optimization (Nonsmooth Optimization )

Citation: The 10th International Conference on Learning Representations

Download: [PDF]

Entry Submitted: 12/04/2021
Entry Accepted: 12/05/2021
Entry Last Modified: 03/15/2022

Modify/Update this entry


  Visitors Authors More about us Links
  Subscribe, Unsubscribe
Digest Archive
Search, Browse the Repository

 

Submit
Update
Policies
Coordinator's Board
Classification Scheme
Credits
Give us feedback
Optimization Journals, Sites, Societies
Mathematical Optimization Society