Back to Search Start Over

Self-supervised contrastive speaker verification with nearest neighbor positive instances.

Authors :
Liu, Yan
Wei, Li-Fang
Zhang, Chuan-Fei
Zhang, Tian-Hao
Chen, Song-Lu
Yin, Xu-Cheng
Source :
Pattern Recognition Letters. Sep2023, Vol. 173, p17-22. 6p.
Publication Year :
2023

Abstract

• We use nearest neighbor positive instances selected from a dynamic queue to improve the positive instances diversity. • Our self-supervised contrastive training model achieves competitive performance compared to previous models. • Our self-supervised contrastive training model is better than supervised training models in cross-dataset testing. Self-supervised contrastive learning (SSCL) has achieved a great success in speaker verification (SV). All recent works treat within-utterance speaker embeddings (SE) to be positive instances, encouraging them to be as close as possible. However, positive instances from the same utterance have similar channel and related semantic information, which are difficult to distinguish from the speaker features. Moreover, these positive instances can only provide limited variances in a certain speaker. To tackle the above problems, we propose to use nearest neighbor (NN) positive instances for training, which are selected from a dynamic queue. The NN positive instances can provide different channel and semantic information, increasing the variances in a certain speaker. Our proposed method are validated through comprehensive experiments on VoxCeleb and CNCeleb1 datasets, demonstrating its effectiveness in improving both SSCL and fine-tuning results. Additionally, our SSCL model outperforms supervised training model in cross-dataset testing due to the use of massive unlabeled data. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
01678655
Volume :
173
Database :
Academic Search Index
Journal :
Pattern Recognition Letters
Publication Type :
Academic Journal
Accession number :
171311678
Full Text :
https://doi.org/10.1016/j.patrec.2023.07.007