Back to Search Start Over

Operator Learning Meets Numerical Analysis: Improving Neural Networks through Iterative Methods

Authors :
Zappala, Emanuele
Levine, Daniel
He, Sizhuang
Rizvi, Syed
Levy, Sacha
van Dijk, David
Publication Year :
2023

Abstract

Deep neural networks, despite their success in numerous applications, often function without established theoretical foundations. In this paper, we bridge this gap by drawing parallels between deep learning and classical numerical analysis. By framing neural networks as operators with fixed points representing desired solutions, we develop a theoretical framework grounded in iterative methods for operator equations. Under defined conditions, we present convergence proofs based on fixed point theory. We demonstrate that popular architectures, such as diffusion models and AlphaFold, inherently employ iterative operator learning. Empirical assessments highlight that performing iterations through network operators improves performance. We also introduce an iterative graph neural network, PIGN, that further demonstrates benefits of iterations. Our work aims to enhance the understanding of deep learning by merging insights from numerical analysis, potentially guiding the design of future networks with clearer theoretical underpinnings and improved performance.<br />Comment: 27 pages (13+14). 8 Figures and 5 tables. Comments are welcome!

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2310.01618
Document Type :
Working Paper