5 results on '"Kaya, Heysem"'
Search Results
2. Text-based Interpretable Depression Severity Modeling via Symptom Predictions
- Author
-
Van steijn, Floris, Sogancioglu, Gizem, Kaya, Heysem, Sub Social and Affective Computing, Social and Affective Computing, Sub Social and Affective Computing, and Social and Affective Computing
- Subjects
AVEC'19 ,Human-Computer Interaction ,extreme learning machine ,Computer Networks and Communications ,Depression Severity Prediction ,explainability ,Affective Computing ,Computer Vision and Pattern Recognition ,interpretability ,Software - Abstract
Mood disorders in general and depression in particular are common and their impact on individuals and society is high. Roughly 5% of adults worldwide suffer from depression. Commonly, depression diagnosis involves using questionnaires, either clinician-rated or self-reported. Due to the subjectivity in questionnaire methods and high human-related costs involved, there are ongoing efforts to find more objective and easily attainable depression markers. As is the case with recent audio, visual and linguistic applications, state-of-the-art approaches for automated depression severity prediction heavily depend on deep learning and black box modeling without explainability and interpretability considerations. However, for reasons ranging from regulations to understanding the extent and limitations of the model, the clinicians need to understand the decision making process of the model to confidently form their decisions. In this work, we focus on text-based depression severity level prediction on DAIC-WOZ corpus and benefit from PHQ-8 questionnaire items to predict the symptoms as interpretable high level features. We show that using a multi-task regression approach with state-of-the-art text-based features to predict the depression symptoms, it is possible to reach a viable test set Concordance Correlation Coefficient performance comparable to the state-of-the-art systems.
- Published
- 2022
3. Towards using Breathing Features for Multimodal Estimation of Depression Severity
- Author
-
Pessanha, Francisca, Kaya, Heysem, Akdag, Almila, Salah, Albert Ali, Sub Social and Affective Computing, Sub Human-Centered Computing, Social and Affective Computing, Sub Social and Affective Computing, Sub Human-Centered Computing, and Social and Affective Computing
- Subjects
DAIC-WOZ Corpus ,Depression Severity Prediction ,Affective Computing ,Interpretability ,Paralinguistics ,Breathing Analysis - Abstract
Breathing patterns are shown to have strong correlations with emotional states, and hence have promise for automatic mood order prediction and analysis. An essential challenge here is the lack of ground truth for breathing sounds, especially for medical and archival datasets. In this study, we provide a cross-dataset approach for breathing pattern prediction and analyse the contribution of predicted breath signals for the detection of depressive states, using the DAIC-WOZ corpus. We use interpretable features in our models to provide actionable insights. Our experimental evaluation shows that in participants with higher depression scores (as indicated by the eight-item Patient Health Questionnaire, PHQ-8), breathing events tend to be shallow or slow. We furthermore tested linear and non-linear regression models with breathing, linguistic sentiment and conversational features, and show that these simple models outperform the AVEC17 Real-life Depression Recognition Sub-challenge baseline.
- Published
- 2022
4. End-to-End Modeling and Transfer Learning for Audiovisual Emotion Recognition in-the-Wild
- Author
-
Dresvyanskiy, Denis, Ryumina, Elena, Kaya, Heysem, Markitantov, Maxim, Karpov, Alexey, Minker, Wolfgang, Sub Social and Affective Computing, Social and Affective Computing, Sub Social and Affective Computing, and Social and Affective Computing
- Subjects
affective computing ,emotion recognition ,deep learning architectures ,face processing ,multimodal fusion ,multimodal representations ,Human-Computer Interaction ,Multimodal fusion ,Multimodal representations ,Computer Networks and Communications ,Face processing ,Neuroscience (miscellaneous) ,Affective computing ,Emotion recognition ,Deep learning architectures ,Computer Science Applications - Abstract
As emotions play a central role in human communication, automatic emotion recognition has attracted increasing attention in the last two decades. While multimodal systems enjoy high performances on lab-controlled data, they are still far from providing ecological validity on non-lab-controlled, namely “in-the-wild” data. This work investigates audiovisual deep learning approaches to emotion recognition in in-the-wild problem. Inspired by the outstanding performance of end-to-end and transfer learning techniques, we explored the effectiveness of architectures in which a modality-specific Convolutional Neural Network (CNN) is followed by a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) using the AffWild2 dataset under the Affective Behavior Analysis in-the-Wild (ABAW) challenge protocol. We deployed unimodal end-to-end and transfer learning approaches within a multimodal fusion system, which generated final predictions using a weighted score fusion scheme. Exploiting the proposed deep-learning-based multimodal system, we reached a test set challenge performance measure of 48.1% on the ABAW 2020 Facial Expressions challenge, which advances the first-runner-up performance.
- Published
- 2022
5. Modeling Short-Term and Long-Term Dependencies of the Speech Signal for Paralinguistic Emotion Classification
- Author
-
Verkholyak, Oxana, Kaya, Heysem, Karpov, Alexey, Sub Social and Affective Computing, Social and Affective Computing, Sub Social and Affective Computing, and Social and Affective Computing
- Subjects
Speech emotion recognition ,Computer Networks and Communications ,Computer science ,Speech recognition ,Emotion classification ,Context modelling ,feature representation ,Feature selection ,Computational paralinguistics ,lcsh:QA75.5-76.95 ,Artificial Intelligence ,Feature (machine learning) ,Long short-term memory ,context modelling ,Affective computing ,Representation (mathematics) ,affective computing ,computational paralinguistics ,Artificial neural network ,Artificial neural networks ,Applied Mathematics ,Term (time) ,Feature representation ,Recurrent neural network ,Control and Systems Engineering ,speech emotion recognition ,lcsh:Electronic computers. Computer science ,long short-term memory ,artificial neural networks - Abstract
Recently, Speech Emotion Recognition (SER) has become an important research topic of affective computing. It is a difficult problem, where some of the greatest challenges lie in the feature selection and representation tasks. A good feature representation should be able to reflect global trends as well as temporal structure of the signal, since emotions naturally evolve in time; it has become possible with the advent of Recurrent Neural Networks (RNN), which are actively used today for various sequence modeling tasks. This paper proposes a hybrid approach to feature representation, which combines traditionally engineered statistical features with Long Short-Term Memory (LSTM) sequence representation in order to take advantage of both short-term and long-term acoustic characteristics of the signal, therefore capturing not only the general trends but also temporal structure of the signal. The evaluation of the proposed method is done on three publicly available acted emotional speech corpora in three different languages, namely RUSLANA (Russian speech), BUEMODB (Turkish speech) and EMODB (German speech). Compared to the traditional approach, the results of our experiments show an absolute improvement of 2.3% and 2.8% for two out of three databases, and a comparative performance on the third. Therefore, provided enough training data, the proposed method proves effective in modelling emotional content of speech utterances. © 2019 St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences. All rights reserved. Russian Science Foundation, RSF: 18-11-00145 This research is supported by the Russian Science Foundation (project ? 18-11-00145). This research is supported by the Russian Science Foundation (project
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.