Back to Search Start Over

Hierarchical Transformer Encoders for Vietnamese Spelling Correction

Authors :
Tran, Hieu
Dinh, Cuong V.
Phan, Long
Nguyen, Son T.
Publication Year :
2021

Abstract

In this paper, we propose a Hierarchical Transformer model for Vietnamese spelling correction problem. The model consists of multiple Transformer encoders and utilizes both character-level and word-level to detect errors and make corrections. In addition, to facilitate future work in Vietnamese spelling correction tasks, we propose a realistic dataset collected from real-life texts for the problem. We compare our method with other methods and publicly available systems. The proposed method outperforms all of the contemporary methods in terms of recall, precision, and f1-score. A demo version is publicly available.<br />Comment: Accepted by The 34th International Conference on Industrial, Engineering & Other Applications of Applied Intelligent Systems(IEA/AIE 2021)

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2105.13578
Document Type :
Working Paper