Back to Search
Start Over
Information flow in layered networks of non-monotonic units
- Source :
- Journal of Statistical Mechanics: Theory and Experiment
- Publication Year :
- 2015
- Publisher :
- IOP Publishing, 2015.
-
Abstract
- Layered neural networks are feedforward structures that yield robust parallel and distributed pattern recognition. Even though much attention has been paid to pattern retrieval properties in such systems, many aspects of their dynamics are not yet well characterized or understood. In this work we study, at different temperatures, the memory activity and information flows through layered networks in which the elements are the simplest binary odd non-monotonic function. Our results show that, considering a standard Hebbian learning approach, the network information content has its maximum always at the monotonic limit, even though the maximum memory capacity can be found at non-monotonic values for small enough temperatures. Furthermore, we show that such systems exhibit rich macroscopic dynamics, including not only fixed point solutions of its iterative map, but also cyclic and chaotic attractors that also carry information.
- Subjects :
- Statistics and Probability
Theoretical computer science
Artificial neural network
Chaotic
Statistical and Nonlinear Physics
Monotonic function
Function (mathematics)
Fixed point
Topology
Hebbian theory
Attractor
Pattern recognition (psychology)
Statistics, Probability and Uncertainty
Mathematics
Subjects
Details
- ISSN :
- 17425468
- Volume :
- 2015
- Database :
- OpenAIRE
- Journal :
- Journal of Statistical Mechanics: Theory and Experiment
- Accession number :
- edsair.doi.dedup.....ad24c4dfd7c7d41ea6bc5a5e4e92adb0
- Full Text :
- https://doi.org/10.1088/1742-5468/2015/07/p07022