Back to Search
Start Over
On a Sparse Shortcut Topology of Artificial Neural Networks
- Source :
- IEEE Transactions on Artificial Intelligence. 3:595-608
- Publication Year :
- 2022
- Publisher :
- Institute of Electrical and Electronics Engineers (IEEE), 2022.
-
Abstract
- In established network architectures, shortcut connections are often used to take the outputs of earlier layers as additional inputs to later layers. Despite the extraordinary effectiveness of shortcuts, there remain open questions on the mechanism and characteristics. For example, why are shortcuts powerful? Why do shortcuts generalize well? In this paper, we investigate the expressivity and generalizability of a novel sparse shortcut topology. First, we demonstrate that this topology can empower a one-neuron-wide deep network to approximate any univariate continuous function. Then, we present a novel width-bounded universal approximator in contrast to depth-bounded universal approximators and extend the approximation result to a family of equally competent networks. Furthermore, with generalization bound theory, we show that the proposed shortcut topology enjoys excellent generalizability. Finally, we corroborate our theoretical analyses by comparing the proposed topology with popular architectures, including ResNet and DenseNet, on well-known benchmarks and perform a saliency map analysis to interpret the proposed topology. Our work helps enhance the understanding of the role of shortcuts and suggests further opportunities to innovate neural architectures.
- Subjects :
- FOS: Computer and information sciences
Computer Science - Machine Learning
Network architecture
Artificial neural network
Continuous function
Computer science
Generalization
Univariate
Contrast (statistics)
Machine Learning (stat.ML)
Topology (electrical circuits)
Topology
Machine Learning (cs.LG)
Statistics - Machine Learning
Generalizability theory
Subjects
Details
- ISSN :
- 26914581
- Volume :
- 3
- Database :
- OpenAIRE
- Journal :
- IEEE Transactions on Artificial Intelligence
- Accession number :
- edsair.doi.dedup.....f1f868f14c04bc3c7472d18b8057d2b3
- Full Text :
- https://doi.org/10.1109/tai.2021.3128132