Back to Search Start Over

Deriving Ground Truth Labels for Regression Problems Using Annotator Precision †.

Authors :
Johnston, Benjamin
de Chazal, Philip
Source :
Applied Sciences (2076-3417); Aug2023, Vol. 13 Issue 16, p9130, 15p
Publication Year :
2023

Abstract

When training machine learning models with practical applications, a quality ground truth dataset is critical. Unlike in classification problems, there is currently no effective method for determining a single ground truth value or landmark from a set of annotations in regression problems. We propose a novel method for deriving ground truth labels in regression problems that considers the performance and precision of individual annotators when identifying each label separately. In contrast to the commonly accepted method of computing the global mean, our method does not assume each annotator to be equally capable of completing the specified task, but rather ensures that higher-performing annotators have a greater contribution to the final result. The ground truth selection method described within this paper provides a means of improving the quality of input data for machine learning model development by removing lower-quality labels. In this study, we objectively demonstrate the improved performance by applying the method to a simulated dataset where a canonical ground truth position can be known, as well as to a sample of data collected from crowd-sourced labels. [ABSTRACT FROM AUTHOR]

Subjects

Subjects :
MACHINE learning
CROWDSOURCING

Details

Language :
English
ISSN :
20763417
Volume :
13
Issue :
16
Database :
Complementary Index
Journal :
Applied Sciences (2076-3417)
Publication Type :
Academic Journal
Accession number :
170711261
Full Text :
https://doi.org/10.3390/app13169130