Back to Search
Start Over
Max-Pooling Loss Training of Long Short-Term Memory Networks for Small-Footprint Keyword Spotting
- Source :
- SLT
- Publication Year :
- 2017
-
Abstract
- We propose a max-pooling based loss function for training Long Short-Term Memory (LSTM) networks for small-footprint keyword spotting (KWS), with low CPU, memory, and latency requirements. The max-pooling loss training can be further guided by initializing with a cross-entropy loss trained network. A posterior smoothing based evaluation approach is employed to measure keyword spotting performance. Our experimental results show that LSTM models trained using cross-entropy loss or max-pooling loss outperform a cross-entropy loss trained baseline feed-forward Deep Neural Network (DNN). In addition, max-pooling loss trained LSTM with randomly initialized network performs better compared to cross-entropy loss trained LSTM. Finally, the max-pooling loss trained LSTM initialized with a cross-entropy pre-trained network shows the best performance, which yields $67.6\%$ relative reduction compared to baseline feed-forward DNN in Area Under the Curve (AUC) measure.
- Subjects :
- FOS: Computer and information sciences
Computer Science - Computation and Language
Artificial neural network
Computer science
Speech recognition
education
Initialization
Machine Learning (stat.ML)
Context (language use)
02 engineering and technology
Machine Learning (cs.LG)
Reduction (complexity)
030507 speech-language pathology & audiology
03 medical and health sciences
Computer Science - Learning
Statistics - Machine Learning
Keyword spotting
0202 electrical engineering, electronic engineering, information engineering
020201 artificial intelligence & image processing
Latency (engineering)
0305 other medical science
Hidden Markov model
Computation and Language (cs.CL)
Smoothing
Subjects
Details
- Language :
- English
- Database :
- OpenAIRE
- Journal :
- SLT
- Accession number :
- edsair.doi.dedup.....8d7cde449e9c0813f2feff82c87bea0c