Back to Search Start Over

RankQA: Neural Question Answering with Answer Re-Ranking

Authors :
Bernhard Kratzwald
Anna Eigenmann
Stefan Feuerriegel
Source :
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019), ACL (1)
Publication Year :
2019
Publisher :
arXiv, 2019.

Abstract

The conventional paradigm in neural question answering (QA) for narrative content is limited to a two-stage process: first, relevant text passages are retrieved and, subsequently, a neural network for machine comprehension extracts the likeliest answer. However, both stages are largely isolated in the status quo and, hence, information from the two phases is never properly fused. In contrast, this work proposes RankQA: RankQA extends the conventional two-stage process in neural QA with a third stage that performs an additional answer re-ranking. The re-ranking leverages different features that are directly extracted from the QA pipeline, i.e., a combination of retrieval and comprehension features. While our intentionally simple design allows for an efficient, data-sparse estimation, it nevertheless outperforms more complex QA systems by a significant margin: in fact, RankQA achieves state-of-the-art performance on 3 out of 4 benchmark datasets. Furthermore, its performance is especially superior in settings where the size of the corpus is dynamic. Here the answer re-ranking provides an effective remedy against the underlying noise-information trade-off due to a variable corpus size. As a consequence, RankQA represents a novel, powerful, and thus challenging baseline for future research in content-based QA.<br />Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019)<br />ISBN:978-1-5108-9096-1<br />ISBN:978-1-950737-48-2

Details

ISBN :
978-1-5108-9096-1
978-1-950737-48-2
ISBNs :
9781510890961 and 9781950737482
Database :
OpenAIRE
Journal :
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019), ACL (1)
Accession number :
edsair.doi.dedup.....84768cc964cda4e45c2ce54a8da62520
Full Text :
https://doi.org/10.48550/arxiv.1906.03008