Back to Search Start Over

Shared computational principles for language processing in humans and deep language models.

Authors :
Goldstein A
Zada Z
Buchnik E
Schain M
Price A
Aubrey B
Nastase SA
Feder A
Emanuel D
Cohen A
Jansen A
Gazula H
Choe G
Rao A
Kim C
Casto C
Fanda L
Doyle W
Friedman D
Dugan P
Melloni L
Reichart R
Devore S
Flinker A
Hasenfratz L
Levy O
Hassidim A
Brenner M
Matias Y
Norman KA
Devinsky O
Hasson U
Source :
Nature neuroscience [Nat Neurosci] 2022 Mar; Vol. 25 (3), pp. 369-380. Date of Electronic Publication: 2022 Mar 07.
Publication Year :
2022

Abstract

Departing from traditional linguistic models, advances in deep learning have resulted in a new type of predictive (autoregressive) deep language models (DLMs). Using a self-supervised next-word prediction task, these models generate appropriate linguistic responses in a given context. In the current study, nine participants listened to a 30-min podcast while their brain responses were recorded using electrocorticography (ECoG). We provide empirical evidence that the human brain and autoregressive DLMs share three fundamental computational principles as they process the same natural narrative: (1) both are engaged in continuous next-word prediction before word onset; (2) both match their pre-onset predictions to the incoming word to calculate post-onset surprise; (3) both rely on contextual embeddings to represent words in natural contexts. Together, our findings suggest that autoregressive DLMs provide a new and biologically feasible computational framework for studying the neural basis of language.<br /> (© 2022. The Author(s).)

Details

Language :
English
ISSN :
1546-1726
Volume :
25
Issue :
3
Database :
MEDLINE
Journal :
Nature neuroscience
Publication Type :
Academic Journal
Accession number :
35260860
Full Text :
https://doi.org/10.1038/s41593-022-01026-4