Back to Search
Start Over
Development of a Human Evaluation Framework and Correlation with Automated Metrics for Natural Language Generation of Medical Diagnoses.
- Source :
-
MedRxiv : the preprint server for health sciences [medRxiv] 2024 Apr 09. Date of Electronic Publication: 2024 Apr 09. - Publication Year :
- 2024
-
Abstract
- In the evolving landscape of clinical Natural Language Generation (NLG), assessing abstractive text quality remains challenging, as existing methods often overlook generative task complexities. This work aimed to examine the current state of automated evaluation metrics in NLG in healthcare. To have a robust and well-validated baseline with which to examine the alignment of these metrics, we created a comprehensive human evaluation framework. Employing ChatGPT-3.5-turbo generative output, we correlated human judgments with each metric. None of the metrics demonstrated high alignment; however, the SapBERT score-a Unified Medical Language System (UMLS)- showed the best results. This underscores the importance of incorporating domain-specific knowledge into evaluation efforts. Our work reveals the deficiency in quality evaluations for generated text and introduces our comprehensive human evaluation framework as a baseline. Future efforts should prioritize integrating medical knowledge databases to enhance the alignment of automated metrics, particularly focusing on refining the SapBERT score for improved assessments.
Details
- Language :
- English
- Database :
- MEDLINE
- Journal :
- MedRxiv : the preprint server for health sciences
- Accession number :
- 38562730
- Full Text :
- https://doi.org/10.1101/2024.03.20.24304620