Back to Search Start Over

A law of data separation in deep learning.

Authors :
Hangfeng He
Su, Weijie J.
Source :
Proceedings of the National Academy of Sciences of the United States of America. 9/5/2023, Vol. 120 Issue 36, p1-8. 23p.
Publication Year :
2023

Abstract

While deep learning has enabled significant advances in many areas of science, its blackbox nature hinders architecture design for future artificial intelligence applications and interpretation for high-stakes decision-makings. We addressed this issue by studying the fundamental question of how deep neural networks process data in the intermediate layers. Our finding is a simple and quantitative law that governs how deep neural networks separate data according to class membership throughout all layers for classification. This law shows that each layer improves data separation at a constant geometric rate, and its emergence is observed in a collection of network architectures and datasets during training. This law offers practical guidelines for designing architectures, improving model robustness and out-of-sample performance, as well as interpreting the predictions. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00278424
Volume :
120
Issue :
36
Database :
Academic Search Index
Journal :
Proceedings of the National Academy of Sciences of the United States of America
Publication Type :
Academic Journal
Accession number :
171865189
Full Text :
https://doi.org/10.1073/pnas.2221704120