Back to Search
Start Over
Self-supervised contrastive speaker verification with nearest neighbor positive instances.
- Source :
-
Pattern Recognition Letters . Sep2023, Vol. 173, p17-22. 6p. - Publication Year :
- 2023
-
Abstract
- • We use nearest neighbor positive instances selected from a dynamic queue to improve the positive instances diversity. • Our self-supervised contrastive training model achieves competitive performance compared to previous models. • Our self-supervised contrastive training model is better than supervised training models in cross-dataset testing. Self-supervised contrastive learning (SSCL) has achieved a great success in speaker verification (SV). All recent works treat within-utterance speaker embeddings (SE) to be positive instances, encouraging them to be as close as possible. However, positive instances from the same utterance have similar channel and related semantic information, which are difficult to distinguish from the speaker features. Moreover, these positive instances can only provide limited variances in a certain speaker. To tackle the above problems, we propose to use nearest neighbor (NN) positive instances for training, which are selected from a dynamic queue. The NN positive instances can provide different channel and semantic information, increasing the variances in a certain speaker. Our proposed method are validated through comprehensive experiments on VoxCeleb and CNCeleb1 datasets, demonstrating its effectiveness in improving both SSCL and fine-tuning results. Additionally, our SSCL model outperforms supervised training model in cross-dataset testing due to the use of massive unlabeled data. [ABSTRACT FROM AUTHOR]
- Subjects :
- *SUPERVISED learning
*SEMANTICS
*SUCCESS
Subjects
Details
- Language :
- English
- ISSN :
- 01678655
- Volume :
- 173
- Database :
- Academic Search Index
- Journal :
- Pattern Recognition Letters
- Publication Type :
- Academic Journal
- Accession number :
- 171311678
- Full Text :
- https://doi.org/10.1016/j.patrec.2023.07.007