1. Images of the unseen: Extrapolating visual representations for abstract and concrete words in a data-driven computational model
- Author
-
Fritz Günther, Alessandra Vergallito, Marco Alessandro Petilli, Marco Marelli, Gunther, F, Petilli, M, Vergallito, A, and Marelli, M
- Subjects
Text corpus ,Computer science ,Concept Formation ,media_common.quotation_subject ,Grounded Cognition ,Experimental and Cognitive Psychology ,Concreteness ,Cognition ,Deep Learning ,Arts and Humanities (miscellaneous) ,Perception ,Distributional Semantic ,Developmental and Educational Psychology ,Humans ,Learning ,Child ,Concretene ,Language ,media_common ,Representation (systemics) ,Abstract and concrete ,General Medicine ,Semantics ,Abstract concept ,Abstract word ,Distributional semantics ,Direct experience ,Language Experience Approach ,Cognitive psychology - Abstract
Theories of grounded cognition assume that conceptual representations are grounded in sensorimotor experience. However, abstract concepts such as jealousy or childhood have no directly associated referents with which such sensorimotor experience can be made; therefore, the grounding of abstract concepts has long been a topic of debate. Here, we propose (a) that systematic relations exist between semantic representations learned from language on the one hand and perceptual experience on the other hand, (b) that these relations can be learned in a bottom-up fashion, and (c) that it is possible to extrapolate from this learning experience to predict expected perceptual representations for words even where direct experience is missing. To test this, we implement a data-driven computational model that is trained to map language-based representations (obtained from text corpora, representing language experience) onto vision-based representations (obtained from an image database, representing perceptual experience), and apply its mapping function onto language-based representations for abstract and concrete words outside the training set. In three experiments, we present participants with these words, accompanied by two images: the image predicted by the model and a random control image. Results show that participants’ judgements were in line with model predictions even for the most abstract words. This preference was stronger for more concrete items and decreased for the more abstract ones. Taken together, our findings have substantial implications in support of the grounding of abstract words, suggesting that we can tap into our previous experience to create possible visual representation we don’t have.
- Published
- 2022
- Full Text
- View/download PDF