Back to Search Start Over

Improving the Validity and Practical Usefulness of AI/ML Evaluations Using an Estimands Framework

Authors :
Binette, Olivier
Reiter, Jerome P.
Publication Year :
2024

Abstract

Commonly, AI or machine learning (ML) models are evaluated on benchmark datasets. This practice supports innovative methodological research, but benchmark performance can be poorly correlated with performance in real-world applications -- a construct validity issue. To improve the validity and practical usefulness of evaluations, we propose using an estimands framework adapted from international clinical trials guidelines. This framework provides a systematic structure for inference and reporting in evaluations, emphasizing the importance of a well-defined estimation target. We illustrate our proposal on examples of commonly used evaluation methodologies - involving cross-validation, clustering evaluation, and LLM benchmarking - that can lead to incorrect rankings of competing models (rank reversals) with high probability, even when performance differences are large. We demonstrate how the estimands framework can help uncover underlying issues, their causes, and potential solutions. Ultimately, we believe this framework can improve the validity of evaluations through better-aligned inference, and help decision-makers and model users interpret reported results more effectively.<br />Comment: 25 pages, 2 figures, 3 tables

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2406.10366
Document Type :
Working Paper