Back to Search
Start Over
Operator Learning Meets Numerical Analysis: Improving Neural Networks through Iterative Methods
- Publication Year :
- 2023
-
Abstract
- Deep neural networks, despite their success in numerous applications, often function without established theoretical foundations. In this paper, we bridge this gap by drawing parallels between deep learning and classical numerical analysis. By framing neural networks as operators with fixed points representing desired solutions, we develop a theoretical framework grounded in iterative methods for operator equations. Under defined conditions, we present convergence proofs based on fixed point theory. We demonstrate that popular architectures, such as diffusion models and AlphaFold, inherently employ iterative operator learning. Empirical assessments highlight that performing iterations through network operators improves performance. We also introduce an iterative graph neural network, PIGN, that further demonstrates benefits of iterations. Our work aims to enhance the understanding of deep learning by merging insights from numerical analysis, potentially guiding the design of future networks with clearer theoretical underpinnings and improved performance.<br />Comment: 27 pages (13+14). 8 Figures and 5 tables. Comments are welcome!
- Subjects :
- Computer Science - Machine Learning
Mathematics - Numerical Analysis
Subjects
Details
- Database :
- arXiv
- Publication Type :
- Report
- Accession number :
- edsarx.2310.01618
- Document Type :
- Working Paper