Back to Search Start Over

Unconstrained Dysfluency Modeling for Dysfluent Speech Transcription and Detection

Authors :
Lian, Jiachen
Feng, Carly
Farooqi, Naasir
Li, Steve
Kashyap, Anshul
Cho, Cheol Jun
Wu, Peter
Netzorg, Robbie
Li, Tingle
Anumanchipalli, Gopala Krishna
Publication Year :
2023

Abstract

Dysfluent speech modeling requires time-accurate and silence-aware transcription at both the word-level and phonetic-level. However, current research in dysfluency modeling primarily focuses on either transcription or detection, and the performance of each aspect remains limited. In this work, we present an unconstrained dysfluency modeling (UDM) approach that addresses both transcription and detection in an automatic and hierarchical manner. UDM eliminates the need for extensive manual annotation by providing a comprehensive solution. Furthermore, we introduce a simulated dysfluent dataset called VCTK++ to enhance the capabilities of UDM in phonetic transcription. Our experimental results demonstrate the effectiveness and robustness of our proposed methods in both transcription and detection tasks.<br />Comment: 2023 ASRU

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2312.12810
Document Type :
Working Paper