Back to Search Start Over

Linking Neural Collapse and L2 Normalization with Improved Out-of-Distribution Detection in Deep Neural Networks

Authors :
Haas, Jarrod
Yolland, William
Rabus, Bernhard
Publication Year :
2022

Abstract

We propose a simple modification to standard ResNet architectures--L2 normalization over feature space--that substantially improves out-of-distribution (OoD) performance on the previously proposed Deep Deterministic Uncertainty (DDU) benchmark. We show that this change also induces early Neural Collapse (NC), an effect linked to better OoD performance. Our method achieves comparable or superior OoD detection scores and classification accuracy in a small fraction of the training time of the benchmark. Additionally, it substantially improves worst case OoD performance over multiple, randomly initialized models. Though we do not suggest that NC is the sole mechanism or a comprehensive explanation for OoD behaviour in deep neural networks (DNN), we believe NC's simple mathematical and geometric structure can provide a framework for analysis of this complex phenomenon in future work.<br />Comment: 19 pages, accepted by Transactions of Machine Learning Research, 2022

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2209.08378
Document Type :
Working Paper