Back to Search Start Over

Using Visual Speech Information in Masking Methods for Audio Speaker Separation

Authors :
Khan, Faheem Ullah
Milner, Ben P.
Le Cornu, Thomas
Source :
IEEE-ACM Transactions on Audio, Speech, and Language Processing; October 2018, Vol. 26 Issue: 10 p1742-1754, 13p
Publication Year :
2018

Abstract

This paper examines whether visual speech information can be effective within audio-masking-based speaker separation to improve the quality and intelligibility of the target speech. Two visual-only methods of generating an audio mask for speaker separation are first developed. These use a deep neural network to map the visual speech features to an audio feature space from which both visually derived binary masks and visually derived ratio masks are estimated, before application to the speech mixture. Second, an audio ratio masking method forms a baseline approach for speaker separation which is extended to exploit visual speech information to form audio-visual ratio masks. Speech quality and intelligibility tests are carried out on the visual-only, audio-only, and audio-visual masking methods of speaker separation at mixing levels from -10 to +10 dB. These reveal substantial improvements in the target speech when applying the visual-only and audio-only masks, but with highest performance occurring when combining audio and visual information to create the audio-visual masks.

Details

Language :
English
ISSN :
23299290
Volume :
26
Issue :
10
Database :
Supplemental Index
Journal :
IEEE-ACM Transactions on Audio, Speech, and Language Processing
Publication Type :
Periodical
Accession number :
ejs45923400
Full Text :
https://doi.org/10.1109/TASLP.2018.2835719