Back to Search
Start Over
HSKDLR: Lightweight Lip Reading Method Based on Homogeneous Self-Knowledge Distillation
- Source :
- Jisuanji kexue yu tansuo, Vol 17, Iss 11, Pp 2689-2702 (2023)
- Publication Year :
- 2023
- Publisher :
- Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press, 2023.
-
Abstract
- In order to solve the problems of low recognition rate and large amount of calculation in lip reading, this paper proposes a lightweight model for lip reading named HSKDLR (homogeneous self-knowledge distillation for lip reading). Firstly, the S-SE (spatial SE)attention module is designed to pay attention to the spatial features of the lip image, which can construct the i-Ghost Bottleneck (improved Ghost Bottleneck) module to extract the channel features and spatial features of the lip image, thereby improving the accuracy of the lip language recognition model. Secondly, a lip reading model is built based on i-Ghost Bottleneck, which reduces the model computation by optimizing the combination of bottleneck structures to a certain extent. Then, in order to improve the accuracy of the model and reduce time consumption, a model optimization method of the homogeneous self-knowledge distillation (HSKD) is proposed. Finally, this paper employs the HSKD to train the lip reading model and verify its recognition performance. And the experimental results show that HSKDLR has higher recognition accuracy and lower computational complexity than the compared methods. The accuracy of the proposed method on LRW dataset is 87.3%, the floating-point number computation is as low as 2.564 GFLOPs, and the parameter quantity is as low as 3.8723×107. Moreover, HSKD can be applied to most lip reading models to improve recognition accuracy effectively and reduce training time.
Details
- Language :
- Chinese
- ISSN :
- 16739418
- Volume :
- 17
- Issue :
- 11
- Database :
- Directory of Open Access Journals
- Journal :
- Jisuanji kexue yu tansuo
- Publication Type :
- Academic Journal
- Accession number :
- edsdoj.9954e5d7425c477d9761efe66c82ae59
- Document Type :
- article
- Full Text :
- https://doi.org/10.3778/j.issn.1673-9418.2208032