Back to Search
Start Over
Modified self-training based statistical models for image classification and speaker identification
- Source :
- International Journal of Speech Technology. 24:1007-1015
- Publication Year :
- 2021
- Publisher :
- Springer Science and Business Media LLC, 2021.
-
Abstract
- Building a high precision statistical model requires ample amounts of supervised (labeled) data to train the models. In certain domains, it is difficult to acquire large amounts of labeled data, especially applications involving images, speech and video data. At the same time, lots of unlabeled data is available in such applications. Self-training is one of the semi-supervised approaches that enables the use of vast unlabeled data to boost the efficiency of the model along with minimal labeled data. In this work, we propose a variant of the self-training approach that embraces soft labeling of unlabeled examples rather than the hard labeling used in conventional self-training. As our work focuses on image and speaker recognition tasks, Gaussian Mixture Model (GMM) based Bayesian classifier is used as a wrapper in the self-training approach. Our experimental studies on STL10, CIFAR10, MIT (image recognition task) and NIST (speaker recognition task) benchmark datasets indicate that the proposed modified self-training approach offers enhanced efficiency over conventional self-training.
- Subjects :
- Linguistics and Language
Contextual image classification
business.industry
Computer science
Statistical model
Pattern recognition
Speaker recognition
Mixture model
Language and Linguistics
Human-Computer Interaction
Task (computing)
Naive Bayes classifier
ComputingMethodologies_PATTERNRECOGNITION
Benchmark (computing)
NIST
Computer Vision and Pattern Recognition
Artificial intelligence
business
Software
Subjects
Details
- ISSN :
- 15728110 and 13812416
- Volume :
- 24
- Database :
- OpenAIRE
- Journal :
- International Journal of Speech Technology
- Accession number :
- edsair.doi...........2d05c3b39cb016fc1519c906e40c41a3