Back to Search Start Over

What do you mean, BERT? Assessing BERT as a Distributional Semantics Model

Authors :
Mickus, Timothee
Paperno, Denis
Constant, Mathieu
van Deemter, Kees
Source :
Proceedings of the Society for Computation in Linguistics: Vol. 3 (2020), Article 34
Publication Year :
2019

Abstract

Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen as an extension of previous noncontextual distributional semantic models. In this work, we focus on BERT, a deep neural network that produces contextualized embeddings and has set the state-of-the-art in several semantic tasks, and study the semantic coherence of its embedding space. While showing a tendency towards coherence, BERT does not fully live up to the natural expectations for a semantic vector space. In particular, we find that the position of the sentence in which a word occurs, while having no meaning correlates, leaves a noticeable trace on the word embeddings and disturbs similarity relationships.

Details

Database :
arXiv
Journal :
Proceedings of the Society for Computation in Linguistics: Vol. 3 (2020), Article 34
Publication Type :
Report
Accession number :
edsarx.1911.05758
Document Type :
Working Paper
Full Text :
https://doi.org/10.7275/t778-ja71