Back to Search
Start Over
What do you mean, BERT? Assessing BERT as a Distributional Semantics Model
- Source :
- Proceedings of the Society for Computation in Linguistics: Vol. 3 (2020), Article 34
- Publication Year :
- 2019
-
Abstract
- Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen as an extension of previous noncontextual distributional semantic models. In this work, we focus on BERT, a deep neural network that produces contextualized embeddings and has set the state-of-the-art in several semantic tasks, and study the semantic coherence of its embedding space. While showing a tendency towards coherence, BERT does not fully live up to the natural expectations for a semantic vector space. In particular, we find that the position of the sentence in which a word occurs, while having no meaning correlates, leaves a noticeable trace on the word embeddings and disturbs similarity relationships.
- Subjects :
- Computer Science - Computation and Language
Subjects
Details
- Database :
- arXiv
- Journal :
- Proceedings of the Society for Computation in Linguistics: Vol. 3 (2020), Article 34
- Publication Type :
- Report
- Accession number :
- edsarx.1911.05758
- Document Type :
- Working Paper
- Full Text :
- https://doi.org/10.7275/t778-ja71