Back to Search Start Over

A note about why deep learning is deep: A discontinuous approximation perspective.

Authors :
Li, Yongxin
Qi, Haobo
Wang, Hansheng
Source :
Stat. Dec2024, Vol. 13 Issue 1, p1-12. 12p.
Publication Year :
2024

Abstract

Deep learning has achieved unprecedented success in recent years. This approach essentially uses the composition of nonlinear functions to model the complex relationship between input features and output labels. However, a comprehensive theoretical understanding of why the hierarchical layered structure can exhibit superior expressive power is still lacking. In this paper, we provide an explanation for this phenomenon by measuring the approximation efficiency of neural networks with respect to discontinuous target functions. We focus on deep neural networks with rectified linear unit (ReLU) activation functions. We find that to achieve the same degree of approximation accuracy, the number of neurons required by a single‐hidden‐layer (SHL) network is exponentially greater than that required by a multi‐hidden‐layer (MHL) network. In practice, discontinuous points tend to contain highly valuable information (i.e., edges in image classification). We argue that this may be a very important reason accounting for the impressive performance of deep neural networks. We validate our theory in extensive experiments. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20491573
Volume :
13
Issue :
1
Database :
Academic Search Index
Journal :
Stat
Publication Type :
Academic Journal
Accession number :
175946311
Full Text :
https://doi.org/10.1002/sta4.654