Back to Search
Start Over
Using Visual Speech Information in Masking Methods for Audio Speaker Separation
- Source :
- IEEE-ACM Transactions on Audio, Speech, and Language Processing; October 2018, Vol. 26 Issue: 10 p1742-1754, 13p
- Publication Year :
- 2018
-
Abstract
- This paper examines whether visual speech information can be effective within audio-masking-based speaker separation to improve the quality and intelligibility of the target speech. Two visual-only methods of generating an audio mask for speaker separation are first developed. These use a deep neural network to map the visual speech features to an audio feature space from which both visually derived binary masks and visually derived ratio masks are estimated, before application to the speech mixture. Second, an audio ratio masking method forms a baseline approach for speaker separation which is extended to exploit visual speech information to form audio-visual ratio masks. Speech quality and intelligibility tests are carried out on the visual-only, audio-only, and audio-visual masking methods of speaker separation at mixing levels from -10 to +10 dB. These reveal substantial improvements in the target speech when applying the visual-only and audio-only masks, but with highest performance occurring when combining audio and visual information to create the audio-visual masks.
Details
- Language :
- English
- ISSN :
- 23299290
- Volume :
- 26
- Issue :
- 10
- Database :
- Supplemental Index
- Journal :
- IEEE-ACM Transactions on Audio, Speech, and Language Processing
- Publication Type :
- Periodical
- Accession number :
- ejs45923400
- Full Text :
- https://doi.org/10.1109/TASLP.2018.2835719