1. 基于 Transformer的多编码器端到端语音识别.
- Author
-
鹿江飞,孙占全
- Abstract
local feature information at shallow layers. To solve this problem, this study proposes a method using multiple encoders to improve the ability of speech feature extraction. An additional convolutional encoder branch is added to strengthen the capture of local feature information, make up for the neglect of local feature information in shallow Transformer, and effectively realize the integration of global and local dependencies of andio feature se quences. In other words, a multi-encoder model hased on Transformer is proposed. Experiments on the open source Chinese Mandarin data set Aishell 1 show that without an external language model, the proposed Transformer-based multi-encoder model has a relative reduction of 4.00% in character error rate when compared with the Transformer model. On the internal non public Shanghainese dialect data set, the performance improve ment of the proposed model is more obvious, and the character error rate is reduced by 48.24% from 19.92% to 10. 31%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF