Back to Search Start Over

Learning to express reward prediction error-like dopaminergic activity requires plastic representations of time.

Authors :
Cone, Ian
Clopath, Claudia
Shouval, Harel Z.
Source :
Nature Communications; 7/12/2024, Vol. 15 Issue 1, p1-17, 17p
Publication Year :
2024

Abstract

The dominant theoretical framework to account for reinforcement learning in the brain is temporal difference learning (TD) learning, whereby certain units signal reward prediction errors (RPE). The TD algorithm has been traditionally mapped onto the dopaminergic system, as firing properties of dopamine neurons can resemble RPEs. However, certain predictions of TD learning are inconsistent with experimental results, and previous implementations of the algorithm have made unscalable assumptions regarding stimulus-specific fixed temporal bases. We propose an alternate framework to describe dopamine signaling in the brain, FLEX (Flexibly Learned Errors in Expected Reward). In FLEX, dopamine release is similar, but not identical to RPE, leading to predictions that contrast to those of TD. While FLEX itself is a general theoretical framework, we describe a specific, biophysically plausible implementation, the results of which are consistent with a preponderance of both existing and reanalyzed experimental data. Reinforcement learning is essential for survival. In this paper, the authors explain why current machine learning models are hard to implement biologically, propose a biologically plausible framework, and show that it agrees with experiments. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20411723
Volume :
15
Issue :
1
Database :
Complementary Index
Journal :
Nature Communications
Publication Type :
Academic Journal
Accession number :
178460283
Full Text :
https://doi.org/10.1038/s41467-024-50205-3