Back to Search
Start Over
Deep Contextualized Self-training for Low Resource Dependency Parsing
- Source :
- Transactions of the Association for Computational Linguistics, Vol 7, Pp 695-713 (2019)
- Publication Year :
- 2019
- Publisher :
- The MIT Press, 2019.
-
Abstract
- Neural dependency parsing has proven very effective, achieving state-of-the-art results on numerous domains and languages. Unfortunately, it requires large amounts of labeled data, that is costly and laborious to create. In this paper we propose a self-training algorithm that alleviates this annotation bottleneck by training a parser on its own output. Our Deep Contextualized Self-training (DCST) algorithm utilizes representation models trained on sequence labeling tasks that are derived from the parser's output when applied to unlabeled data, and integrates these models with the base parser through a gating mechanism. We conduct experiments across multiple languages, both in low resource in-domain and in cross-domain setups, and demonstrate that DCST substantially outperforms traditional self-training as well as recent semi-supervised training methods.<br />Comment: Accepted to TACL in September 2019
- Subjects :
- FOS: Computer and information sciences
Computer Science - Machine Learning
Linguistics and Language
Computer Science - Computation and Language
Low resource
business.industry
Computer science
Communication
lcsh:P98-98.5
computer.software_genre
Machine Learning (cs.LG)
Computer Science Applications
Human-Computer Interaction
Artificial Intelligence
Dependency grammar
Labeled data
Artificial intelligence
lcsh:Computational linguistics. Natural language processing
business
Computation and Language (cs.CL)
computer
Self training
Natural language processing
Subjects
Details
- Language :
- English
- Volume :
- 7
- Database :
- OpenAIRE
- Journal :
- Transactions of the Association for Computational Linguistics
- Accession number :
- edsair.doi.dedup.....f35759ac0b1255a6eadb18ad10ab853e
- Full Text :
- https://doi.org/10.1162/tacl_a_00294