Back to Search Start Over

Ensemble Distilling Pretrained Language Models for Machine Translation Quality Estimation

Authors :
Jin An Xu
Yufeng Chen
Hui Di
Kazushige Ouchi
Hui Huang
Source :
Natural Language Processing and Chinese Computing ISBN: 9783030604561, NLPCC (2)
Publication Year :
2020
Publisher :
Springer International Publishing, 2020.

Abstract

Machine translation quality estimation (Quality Estimation, QE) aims to evaluate the quality of machine translation automatically without golden reference. QE can be implemented on different granularities, thus to give an estimation for different aspects of machines translation output. In this paper, we propose an effective method to utilize pretrained language models to improve the performance of QE. Our model combines two popular pretrained models, which are Bert and XLM, to create a very strong baseline for both sentence-level and word-level QE. We also propose a simple yet effective strategy, ensemble distillation, to further improve the accuracy of QE system. Ensemble distillation can integrate different knowledge from multiple models into one model, and strengthen each single model by a large margin. We evaluate our system on CCMT2019 Chinese-English and English-Chinese QE dataset, which contains word-level and sentence-level subtasks. Experiment results show our model surpasses previous models to a large extend, demonstrating the effectiveness of our proposed method.

Details

ISBN :
978-3-030-60456-1
ISBNs :
9783030604561
Database :
OpenAIRE
Journal :
Natural Language Processing and Chinese Computing ISBN: 9783030604561, NLPCC (2)
Accession number :
edsair.doi...........a5d3cf74724f19ee5882005bba912db7
Full Text :
https://doi.org/10.1007/978-3-030-60457-8_19