Back to Search Start Over

Robust Multimodal Learning for Ophthalmic Disease Grading via Disentangled Representation

Authors :
Wang, Xinkun
Wang, Yifang
Liang, Senwei
Tang, Feilong
Liu, Chengzhi
Hu, Ming
Hu, Chao
He, Junjun
Ge, Zongyuan
Razzak, Imran
Publication Year :
2025

Abstract

This paper discusses how ophthalmologists often rely on multimodal data to improve diagnostic accuracy. However, complete multimodal data is rare in real-world applications due to a lack of medical equipment and concerns about data privacy. Traditional deep learning methods typically address these issues by learning representations in latent space. However, the paper highlights two key limitations of these approaches: (i) Task-irrelevant redundant information (e.g., numerous slices) in complex modalities leads to significant redundancy in latent space representations. (ii) Overlapping multimodal representations make it difficult to extract unique features for each modality. To overcome these challenges, the authors propose the Essence-Point and Disentangle Representation Learning (EDRL) strategy, which integrates a self-distillation mechanism into an end-to-end framework to enhance feature selection and disentanglement for more robust multimodal learning. Specifically, the Essence-Point Representation Learning module selects discriminative features that improve disease grading performance. The Disentangled Representation Learning module separates multimodal data into modality-common and modality-unique representations, reducing feature entanglement and enhancing both robustness and interpretability in ophthalmic disease diagnosis. Experiments on multimodal ophthalmology datasets show that the proposed EDRL strategy significantly outperforms current state-of-the-art methods.<br />Comment: 10pages

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2503.05319
Document Type :
Working Paper