Back to Search Start Over

Topological Properties of the Set of Functions Generated by Neural Networks of Fixed Size.

Authors :
Petersen, Philipp
Raslan, Mones
Voigtlaender, Felix
Source :
Foundations of Computational Mathematics. Apr2021, Vol. 21 Issue 2, p375-444. 70p.
Publication Year :
2021

Abstract

We analyze the topological properties of the set of functions that can be implemented by neural networks of a fixed size. Surprisingly, this set has many undesirable properties. It is highly non-convex, except possibly for a few exotic activation functions. Moreover, the set is not closed with respect to L p -norms, 0 < p < ∞ , for all practically used activation functions, and also not closed with respect to the L ∞ -norm for all practically used activation functions except for the ReLU and the parametric ReLU. Finally, the function that maps a family of weights to the function computed by the associated network is not inverse stable for every practically used activation function. In other words, if f 1 , f 2 are two functions realized by neural networks and if f 1 , f 2 are close in the sense that ‖ f 1 - f 2 ‖ L ∞ ≤ ε for ε > 0 , it is, regardless of the size of ε , usually not possible to find weights w 1 , w 2 close together such that each f i is realized by a neural network with weights w i . Overall, our findings identify potential causes for issues in the training procedure of deep learning such as no guaranteed convergence, explosion of parameters, and slow convergence. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
16153375
Volume :
21
Issue :
2
Database :
Academic Search Index
Journal :
Foundations of Computational Mathematics
Publication Type :
Academic Journal
Accession number :
149789482
Full Text :
https://doi.org/10.1007/s10208-020-09461-0