Back to Search Start Over

How NOT To Evaluate Your Dialogue System: An Empirical Study of Unsupervised Evaluation Metrics for Dialogue Response Generation

Authors :
Liu, Chia-Wei
Lowe, Ryan
Serban, Iulian V.
Noseworthy, Michael
Charlin, Laurent
Pineau, Joelle
Publication Year :
2016

Abstract

We investigate evaluation metrics for dialogue response generation systems where supervised labels, such as task completion, are not available. Recent works in response generation have adopted metrics from machine translation to compare a model's generated response to a single target response. We show that these metrics correlate very weakly with human judgements in the non-technical Twitter domain, and not at all in the technical Ubuntu domain. We provide quantitative and qualitative results highlighting specific weaknesses in existing metrics, and provide recommendations for future development of better automatic evaluation metrics for dialogue systems.<br />Comment: First 4 authors had equal contribution. 13 pages, 5 tables, 6 figures. EMNLP 2016

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.1603.08023
Document Type :
Working Paper