Back to Search Start Over

Dimensionality and ramping: Signatures of sentence integration in the dynamics of brains and deep language models

Authors :
Théo Desbordes
Yair Lakretz
Valérie Chanoine
Maxime Oquab
Jean-Michel Badier
Agnès Trébuchon
Romain Carron
Christian-G. Bénar
Stanislas Dehaene
Jean-Rémi King
Neuroimagerie cognitive - Psychologie cognitive expérimentale (UNICOG-U992)
Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Université Paris-Saclay
Laboratoire Parole et Langage (LPL)
Aix Marseille Université (AMU)-Centre National de la Recherche Scientifique (CNRS)
Meta AI Research [Paris]
Meta AI
Institut de Neurosciences des Systèmes (INS)
Aix Marseille Université (AMU)-Institut National de la Santé et de la Recherche Médicale (INSERM)
Hôpital de la Timone [CHU - APHM] (TIMONE)
Université Paris sciences et lettres (PSL)
ANR-16-CONV-0002,ILCB,ILCB: Institute of Language Communication and the Brain(2016)
Source :
Journal of Neuroscience, Journal of Neuroscience, 2023, pp.JN-RM-1163-22. ⟨10.1523/JNEUROSCI.1163-22.2023⟩
Publication Year :
2023
Publisher :
HAL CCSD, 2023.

Abstract

A sentence is more than the sum of its words: its meaning depends on how they combine with one another. The brain mechanisms underlying such semantic composition remain poorly understood. To shed light on the neural vector code underlying semantic composition, we introduce two hypotheses: First, the intrinsic dimensionality of the space of neural representations should increase as a sentence unfolds, paralleling the growing complexity of its semantic representation, and second, this progressive integration should be reflected in ramping and sentence-final signals. To test these predictions, we designed a dataset of closely matched normal and Jabberwocky sentences (composed of meaningless pseudo words) and displayed them to deep language models and to 11 human participants (5 men and 6 women) monitored with simultaneous magneto-encephalography and intracranial electro-encephalography. In both deep language models and electrophysiological data, we found that representational dimensionality was higher for meaningful sentences than Jabberwocky. Furthermore, multivariate decoding of normal versus Jabberwocky confirmed three dynamic patterns: (i) a phasic pattern following each word, peaking in temporal and parietal areas, (ii) a ramping pattern, characteristic of bilateral inferior and middle frontal gyri, and (iii) a sentence-final pattern in left superior frontal gyrus and right orbitofrontal cortex. These results provide a first glimpse into the neural geometry of semantic integration and constrain the search for a neural code of linguistic composition.Significance statementStarting from general linguistic concepts, we make two sets of predictions in neural signals evoked by reading multi-word sentences. First, the intrinsic dimensionality of the representation should grow with additional meaningful words. Second, the neural dynamics should exhibit signatures of encoding, maintaining, and resolving semantic composition. We successfully validated these hypotheses in deep Neural Language Models, artificial neural networks trained on text and performing very well on many Natural Language Processing tasks. Then, using a unique combination of magnetoencephalography and intracranial electrodes, we recorded high-resolution brain data from human participants while they read a controlled set of sentences. Time-resolved dimensionality analysis showed increasing dimensionality with meaning, and multivariate decoding allowed us to isolate the three dynamical patterns we had hypothesized.

Details

Language :
English
ISSN :
02706474 and 15292401
Database :
OpenAIRE
Journal :
Journal of Neuroscience, Journal of Neuroscience, 2023, pp.JN-RM-1163-22. ⟨10.1523/JNEUROSCI.1163-22.2023⟩
Accession number :
edsair.doi.dedup.....0f1054b2081aeef9990a1ee1eb462992