1. Denoised Bottleneck Features From Deep Autoencoders for Telephone Conversation Analysis
- Author
-
Richard Dufour, Mohamed Morchid, Killian Janod, Georges Linarès, Renato De Mori, Laboratoire Informatique d'Avignon (LIA), and Avignon Université (AU)-Centre d'Enseignement et de Recherche en Informatique - CERI
- Subjects
Acoustics and Ultrasonics ,Computer science ,media_common.quotation_subject ,Speech recognition ,Feature extraction ,02 engineering and technology ,computer.software_genre ,Bottleneck ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,030507 speech-language pathology & audiology ,03 medical and health sciences ,Transcription (linguistics) ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,Computer Science (miscellaneous) ,Conversation ,Electrical and Electronic Engineering ,ComputingMilieux_MISCELLANEOUS ,media_common ,business.industry ,Speech processing ,Autoencoder ,[INFO.INFO-TT]Computer Science [cs]/Document and Text Processing ,Computational Mathematics ,Conversation analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,0305 other medical science ,business ,computer ,Natural language processing - Abstract
Automatic transcription of spoken documents is affected by automatic transcription errors that are especially frequent when speech is acquired in severe noisy conditions. Automatic speech recognition errors induce errors in the linguistic features used for a variety of natural language processing tasks. Recently, denoisng autoencoders (DAE) and stacked autoencoders (SAE) have been proposed with interesting results for acoustic feature denoising tasks. This paper deals with the recovery of corrupted linguistic features in spoken documents. Solutions based on DAEs and SAEs are considered and evaluated in a spoken conversation analysis task. In order to improve conversation theme classification accuracy, the possibility of combining abstractions obtained from manual and automatic transcription features is considered. As a result, two original representations of highly imperfect spoken documents are introduced. They are based on bottleneck features of a supervised autoencoder that takes advantage of both noisy and clean transcriptions to improve the robustness of error prone representations. Experimental results on a spoken conversation theme identification task show substantial accuracy improvements obtained with the proposed recovery of corrupted features.
- Published
- 2017
- Full Text
- View/download PDF