Back to Search Start Over

Understanding How Pretraining Regularizes Deep Learning Algorithms

Authors :
Yao, Yu
Yu, Baosheng
Gong, Chen
Liu, Tongliang
Source :
IEEE Transactions on Neural Networks and Learning Systems; September 2023, Vol. 34 Issue: 9 p5828-5840, 13p
Publication Year :
2023

Abstract

Deep learning algorithms have led to a series of breakthroughs in computer vision, acoustical signal processing, and others. However, they have only been popularized recently due to the groundbreaking techniques developed for training deep architectures. Understanding the training techniques is important if we want to further improve them. Through extensive experimentation, Erhan et al. (2010) empirically illustrated that unsupervised pretraining has an effect of regularization for deep learning algorithms. However, theoretical justifications for the observation remain elusive. In this article, we provide theoretical supports by analyzing how unsupervised pretraining regularizes deep learning algorithms. Specifically, we interpret deep learning algorithms as the traditional Tikhonov-regularized batch learning algorithms that simultaneously learn predictors in the input feature spaces and the parameters of the neural networks to produce the Tikhonov matrices. We prove that unsupervised pretraining helps in learning meaningful Tikhonov matrices, which will make the deep learning algorithms uniformly stable and the learned predictor will generalize fast w.r.t. the sample size. Unsupervised pretraining, therefore, can be interpreted as to have the function of regularization.

Details

Language :
English
ISSN :
2162237x and 21622388
Volume :
34
Issue :
9
Database :
Supplemental Index
Journal :
IEEE Transactions on Neural Networks and Learning Systems
Publication Type :
Periodical
Accession number :
ejs63863209
Full Text :
https://doi.org/10.1109/TNNLS.2021.3131377