1. Multi-channel spectrograms for speech processing applications using deep learning methods
- Author
-
Juan Camilo Vásquez-Correa, Philipp Klumpp, Juan Rafael Orozco-Arroyave, Maria Schuster, Elmar Nöth, and Tomas Arias-Vergara
- Subjects
Audio signal ,business.industry ,Computer science ,Speech recognition ,Deep learning ,Speech processing ,Convolutional neural network ,Convolution ,030507 speech-language pathology & audiology ,03 medical and health sciences ,0302 clinical medicine ,Recurrent neural network ,Artificial Intelligence ,ddc:000 ,Feature (machine learning) ,Spectrogram ,Computer Vision and Pattern Recognition ,Artificial intelligence ,030223 otorhinolaryngology ,0305 other medical science ,business - Abstract
Time–frequency representations of the speech signals provide dynamic information about how the frequency component changes with time. In order to process this information, deep learning models with convolution layers can be used to obtain feature maps. In many speech processing applications, the time–frequency representations are obtained by applying the short-time Fourier transform and using single-channel input tensors to feed the models. However, this may limit the potential of convolutional networks to learn different representations of the audio signal. In this paper, we propose a methodology to combine three different time–frequency representations of the signals by computing continuous wavelet transform, Mel-spectrograms, and Gammatone spectrograms and combining then into 3D-channel spectrograms to analyze speech in two different applications: (1) automatic detection of speech deficits in cochlear implant users and (2) phoneme class recognition to extract phone-attribute features. For this, two different deep learning-based models are considered: convolutional neural networks and recurrent neural networks with convolution layers.
- Published
- 2020