Back to Search
Start Over
Constraint-Based Regularization of Neural Networks
- Source :
- OPT2020: 12th Annual Workshop on Optimization for Machine Learning, NeurIPS 2020
- Publication Year :
- 2020
-
Abstract
- We propose a method for efficiently incorporating constraints into a stochastic gradient Langevin framework for the training of deep neural networks. Constraints allow direct control of the parameter space of the model. Appropriately designed, they reduce the vanishing/exploding gradient problem, control weight magnitudes and stabilize deep neural networks and thus improve the robustness of training algorithms and the generalization capabilities of the trained neural network. We present examples of constrained training methods motivated by orthogonality preservation for weight matrices and explicit weight normalizations. We describe the methods in the overdamped formulation of Langevin dynamics and the underdamped form, in which momenta help to improve sampling efficiency. The methods are explored in test examples in image classification and natural language processing.<br />Comment: T. Vlaar won best student paper award at OPT2020
- Subjects :
- Computer Science - Machine Learning
Statistics - Machine Learning
Subjects
Details
- Database :
- arXiv
- Journal :
- OPT2020: 12th Annual Workshop on Optimization for Machine Learning, NeurIPS 2020
- Publication Type :
- Report
- Accession number :
- edsarx.2006.10114
- Document Type :
- Working Paper