1. Towards Data-Efficient Modeling for Wake Word Spotting
- Author
-
Shiv Naga Prasad Vitaladevuni, Yuriy Mishchenko, Anish Shah, Yixin Gao, and Spyros Matsoukas
- Subjects
FOS: Computer and information sciences ,Sound (cs.SD) ,Computer Science - Machine Learning ,Computer science ,Speech recognition ,020206 networking & telecommunications ,Speech corpus ,02 engineering and technology ,Semi-supervised learning ,Wake ,Spotting ,Computer Science - Sound ,Machine Learning (cs.LG) ,Audio and Speech Processing (eess.AS) ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
Wake word (WW) spotting is challenging in far-field not only because of the interference in signal transmission but also the complexity in acoustic environment. Traditional WW model training requires large amount of in-domain WW-specific data with substantial human annotations. This prevents the model building in the situation of lacking such data. In this paper we present data-efficient solutions to address the challenges in WW modeling, such as domain-mismatch, noisy conditions, limited annotation, etc. Our proposed system is composed of a multi-condition training pipeline with stratified data augmentation, which improves the model robustness to a variety of predefined acoustic conditions, together with a semi-supervised learning pipeline to extract the WW and adversarial examples from untranscribed speech corpus. Starting from only 10 hours of domain-mismatched WW audio, we are able to enlarge and enrich the training dataset by 20-100 times to capture the complexity in acoustic environments. Our experiments on real user data show that the proposed solutions can achieve comparable performance of a production-grade model by saving 97% of the amount of WW-specific data to collect and 86% of the bandwidth for annotation.
- Published
- 2020
- Full Text
- View/download PDF