Back to Search Start Over

Resonator Networks, 2: Factorization Performance and Capacity Compared to Optimization-Based Methods

Authors :
E. Paxon Frady
Bruno A. Olshausen
Spencer J. Kent
Friedrich T. Sommer
Source :
Neural Computation. 32:2332-2388
Publication Year :
2020
Publisher :
MIT Press, 2020.

Abstract

We develop theoretical foundations of resonator networks, a new type of recurrent neural network introduced in Frady, Kent, Olshausen, and Sommer ( 2020 ), a companion article in this issue, to solve a high-dimensional vector factorization problem arising in Vector Symbolic Architectures. Given a composite vector formed by the Hadamard product between a discrete set of high-dimensional vectors, a resonator network can efficiently decompose the composite into these factors. We compare the performance of resonator networks against optimization-based methods, including Alternating Least Squares and several gradient-based algorithms, showing that resonator networks are superior in several important ways. This advantage is achieved by leveraging a combination of nonlinear dynamics and searching in superposition, by which estimates of the correct solution are formed from a weighted superposition of all possible solutions. While the alternative methods also search in superposition, the dynamics of resonator networks allow them to strike a more effective balance between exploring the solution space and exploiting local information to drive the network toward probable solutions. Resonator networks are not guaranteed to converge, but within a particular regime they almost always do. In exchange for relaxing the guarantee of global convergence, resonator networks are dramatically more effective at finding factorizations than all alternative approaches considered.

Details

ISSN :
1530888X and 08997667
Volume :
32
Database :
OpenAIRE
Journal :
Neural Computation
Accession number :
edsair.doi.dedup.....10bd223dcb1e613bb3033032ecaea196