1. Making Language Model as Small as Possible in Statistical Machine Translation
- Author
-
Jie Hao, Dakun Zhang, Yang Liu, and Jiajun Zhang
- Subjects
Machine translation ,business.industry ,Computer science ,Transfer-based machine translation ,computer.software_genre ,Machine translation software usability ,Example-based machine translation ,Rule-based machine translation ,Cache language model ,Computer-assisted translation ,Language model ,Artificial intelligence ,business ,computer ,Natural language processing - Abstract
As one of the key components, n-gram language model is most frequently used in statistical machine translation. Typically, higher order of the language model leads to better translation performance. However, higher order of the n-gram language model requires much more monolingual training data to avoid data sparseness. Furthermore, the model size increases exponentially when the n-gram order becomes higher and higher. In this paper, we investigate the language model pruning techniques that aim at making the model size as small as possible while keeping the translation quality. According to our investigation, we further propose to replace the higher order n-grams with a low-order cluster-based language model. The extensive experiments show that our method is very effective.
- Published
- 2014
- Full Text
- View/download PDF