1. Neural learning rules for generating flexible predictions and computing the successor representation
- Author
-
Ching Fang, Dmitriy Aronov, LF Abbott, and Emily L Mackevicius
- Subjects
tufted titmouse ,hippocampus ,state-space model ,recurrent neural network ,plasticity ,predictive coding ,Medicine ,Science ,Biology (General) ,QH301-705.5 - Abstract
The predictive nature of the hippocampus is thought to be useful for memory-guided cognitive behaviors. Inspired by the reinforcement learning literature, this notion has been formalized as a predictive map called the successor representation (SR). The SR captures a number of observations about hippocampal activity. However, the algorithm does not provide a neural mechanism for how such representations arise. Here, we show the dynamics of a recurrent neural network naturally calculate the SR when the synaptic weights match the transition probability matrix. Interestingly, the predictive horizon can be flexibly modulated simply by changing the network gain. We derive simple, biologically plausible learning rules to learn the SR in a recurrent network. We test our model with realistic inputs and match hippocampal data recorded during random foraging. Taken together, our results suggest that the SR is more accessible in neural circuits than previously thought and can support a broad range of cognitive functions.
- Published
- 2023
- Full Text
- View/download PDF