1. The different brain areas occupied for integrating information of hierarchical linguistic units: a study based on EEG and TMS
- Author
-
Y. Li, Y. Si, C. Chen, X. Huang, Q. Liu, D. Yao, Kimmo Alho, Y. Qiu, X. Zhang, Z. Cao, S. Gao, C. Pei, F. Li, N. Ding, P. Xu, Pei, Changfu, Qiu, Yuan, Li, Fali, Huang, Xunan, Si, Yajing, Li, Yuqin, Zhang, Xiabing, Chen, Chunli, Liu, Qiang, Cao, Zehong, Ding, Nai, Gao, Shan, Alho, Kimmo, Yao, Dezhong, and Xu, Peng
- Subjects
Modality (human–computer interaction) ,Modalities ,Chinese ,medicine.diagnostic_test ,Computer science ,Cognitive Neuroscience ,CTBS ,Information processing ,linguistic hierarchy ,Experimental Psychology ,Electroencephalography ,Signal ,Linguistics ,Comprehension ,Cellular and Molecular Neuroscience ,transcranial magnetic stimulation ,medicine ,EEG ,audio-visual integration ,Sentence ,1109 Neurosciences, 1701 Psychology, 1702 Cognitive Sciences - Abstract
Human linguistic units are hierarchical, and our brain responds differently when processing linguistic units during sentence comprehension, especially when the modality of the received signal is different (auditory, visual, or audio-visual). However, it is unclear how the brain processes and integrates language information at different linguistic units (words, phrases, and sentences) provided simultaneously in audio and visual modalities. To address the issue, we presented participants with sequences of short Chinese sentences through auditory or visual or combined audio- visual modalities, while electroencephalographic responses were recorded. With a frequency tagging approach, we analyzed the neural representations of basic linguistic units (i.e., characters/monosyllabic words) and higher-level linguistic structures (i.e., phrases and sentences) across the three modalities separately. We found that audio-visual integration occurs at all linguistic units, and the brain areas involved in the integration varied across different linguistic levels. In particular, the integration of sentences activated the local left prefrontal area. Therefore, we used continuous theta-burst stimulation (cTBS) to verify that the left prefrontal cortex plays a vital role in the audio-visual integration of sentence information. Our findings suggest the advantage of bimodal language comprehension at hierarchical stages in language-related information processing and provide evidence for the causal role of the left prefrontal regions in processing information of audio-visual sentences.
- Published
- 2022