Back to Search Start Over

Context-Aware Testing: A New Paradigm for Model Testing with Large Language Models

Authors :
Rauba, Paulius
Seedat, Nabeel
Luyten, Max Ruiz
van der Schaar, Mihaela
Publication Year :
2024

Abstract

The predominant de facto paradigm of testing ML models relies on either using only held-out data to compute aggregate evaluation metrics or by assessing the performance on different subgroups. However, such data-only testing methods operate under the restrictive assumption that the available empirical data is the sole input for testing ML models, disregarding valuable contextual information that could guide model testing. In this paper, we challenge the go-to approach of data-only testing and introduce context-aware testing (CAT) which uses context as an inductive bias to guide the search for meaningful model failures. We instantiate the first CAT system, SMART Testing, which employs large language models to hypothesize relevant and likely failures, which are evaluated on data using a self-falsification mechanism. Through empirical evaluations in diverse settings, we show that SMART automatically identifies more relevant and impactful failures than alternatives, demonstrating the potential of CAT as a testing paradigm.<br />Comment: Presented at the 38th Conference on Neural Information Processing Systems (NeurIPS 2024). *Rauba & Seedat contributed equally

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2410.24005
Document Type :
Working Paper