Back to Search
Start Over
Assessors agreement: A case study across assessor type, payment levels, query variations and relevance dimensions
- Source :
- Experimental IR Meets Multilinguality, Multimodality, and Interaction: 7th International Conference of the CLEF Association, CLEF 2016, Proceedings [Lecture Notes in Computer Science, Volume 9822]
- Publication Year :
- 2016
-
Abstract
- Relevance assessments are the cornerstone of Information Retrieval evaluation. Yet, there is only limited understanding of how assessment disagreement influences the reliability of the evaluation in terms of systems rankings. In this paper we examine the role of assessor type (expert vs. layperson), payment levels (paid vs. unpaid), query variations and relevance dimensions (topicality and understandability) and their influence on system evaluation in the presence of disagreements across assessments obtained in the different settings. The analysis is carried out in the context of the CLEF 2015 eHealth Task 2 collection and shows that disagreements between assessors belonging to the same group have little impact on evaluation. It also shows, however, that assessment disagreement found across settings has major impact on evaluation when topical relevance is considered, while it has no impact when understandability assessments are considered.
Details
- Database :
- OAIster
- Journal :
- Experimental IR Meets Multilinguality, Multimodality, and Interaction: 7th International Conference of the CLEF Association, CLEF 2016, Proceedings [Lecture Notes in Computer Science, Volume 9822]
- Notes :
- application/pdf
- Publication Type :
- Electronic Resource
- Accession number :
- edsoai.on1146607292
- Document Type :
- Electronic Resource