Back to Search Start Over

On the Comparison between Multi-modal and Single-modal Contrastive Learning

Authors :
Huang, Wei
Han, Andi
Chen, Yongqiang
Cao, Yuan
Xu, Zhiqiang
Suzuki, Taiji
Source :
NeurIPS 2024
Publication Year :
2024

Abstract

Multi-modal contrastive learning with language supervision has presented a paradigm shift in modern machine learning. By pre-training on a web-scale dataset, multi-modal contrastive learning can learn high-quality representations that exhibit impressive robustness and transferability. Despite its empirical success, the theoretical understanding is still in its infancy, especially regarding its comparison with single-modal contrastive learning. In this work, we introduce a feature learning theory framework that provides a theoretical foundation for understanding the differences between multi-modal and single-modal contrastive learning. Based on a data generation model consisting of signal and noise, our analysis is performed on a ReLU network trained with the InfoMax objective function. Through a trajectory-based optimization analysis and generalization characterization on downstream tasks, we identify the critical factor, which is the signal-to-noise ratio (SNR), that impacts the generalizability in downstream tasks of both multi-modal and single-modal contrastive learning. Through the cooperation between the two modalities, multi-modal learning can achieve better feature learning, leading to improvements in performance in downstream tasks compared to single-modal learning. Our analysis provides a unified framework that can characterize the optimization and generalization of both single-modal and multi-modal contrastive learning. Empirical experiments on both synthetic and real-world datasets further consolidate our theoretical findings.<br />Comment: 51pages, 1 figure, 1 table

Details

Database :
arXiv
Journal :
NeurIPS 2024
Publication Type :
Report
Accession number :
edsarx.2411.02837
Document Type :
Working Paper