Back to Search Start Over

Disentangling Generative Factors in Natural Language with Discrete Variational Autoencoders

Authors :
Mercatali, Giangiacomo
Freitas, André
Publication Year :
2021

Abstract

The ability of learning disentangled representations represents a major step for interpretable NLP systems as it allows latent linguistic features to be controlled. Most approaches to disentanglement rely on continuous variables, both for images and text. We argue that despite being suitable for image datasets, continuous variables may not be ideal to model features of textual data, due to the fact that most generative factors in text are discrete. We propose a Variational Autoencoder based method which models language features as discrete variables and encourages independence between variables for learning disentangled representations. The proposed model outperforms continuous and discrete baselines on several qualitative and quantitative benchmarks for disentanglement as well as on a text style transfer downstream application.<br />Comment: Findings of EMNLP 2021

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2109.07169
Document Type :
Working Paper