1. Neural architectures for aggregating sequence labels from multiple annotators.
- Author
-
Li, Maolin and Ananiadou, Sophia
- Subjects
- *
ARTIFICIAL neural networks , *MARKOV processes , *VIRTUAL networks , *NATURAL language processing - Abstract
• Explore two neural network-based methods to aggregate noisy sequence labels. • Provide a detailed comparison of different methods: probabilistic graphical models, neural networks and methods that combine graphical and neural network models. • Improve true label prediction performance on three real-world datasets from different domains and tasks. • Present a simple but effective method to learn meaningful annotator embeddings. • Demonstrate our annotator embedding is helpful for studying annotators' reliability and behaviour. Labelled data for training sequence labelling models can be collected from multiple annotators or workers in crowdsourcing. However, these labels could be noisy because of the varying expertise and reliability of annotators. In order to ensure high quality of data, it is crucial to infer the correct labels by aggregating noisy labels. Although label aggregation is a well-studied topic, only a number of studies have investigated how to aggregate sequence labels. Recently, neural network models have attracted research attention for this task. In this paper, we explore two neural network-based methods. The first method combines Hidden Markov Models with networks while also learning distributed representations of annotators (i.e., annotator embedding); the second method combines BiLSTM with autoencoders. The experimental results on three real-world datasets demonstrate the effectiveness of using neural networks for sequence label aggregation. Moreover, our analysis shows that annotators' embeddings not only make our model applicable to real-time applications, but also useful for studying the behaviour of annotators. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF