Back to Search Start Over

DMT: Comprehensive Distillation with Multiple Self-supervised Teachers

Authors :
Liu, Yuang
Wang, Jing
Zhou, Qiang
Wang, Fan
Wang, Jun
Zhang, Wei
Publication Year :
2023

Abstract

Numerous self-supervised learning paradigms, such as contrastive learning and masked image modeling, have been proposed to acquire powerful and general representations from unlabeled data. However, these models are commonly pretrained within their specific framework alone, failing to consider the complementary nature of visual representations. To tackle this issue, we introduce Comprehensive Distillation with Multiple Self-supervised Teachers (DMT) for pretrained model compression, which leverages the strengths of multiple off-the-shelf self-supervised models. Our experimental results on prominent benchmark datasets exhibit that the proposed method significantly surpasses state-of-the-art competitors while retaining favorable efficiency metrics. On classification tasks, our DMT framework utilizing three different self-supervised ViT-Base teachers enhances the performance of both small/tiny models and the base model itself. For dense tasks, DMT elevates the AP/mIoU of standard SSL models on MS-COCO and ADE20K datasets by 4.0%.<br />Comment: ICASSP 2024

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2312.11938
Document Type :
Working Paper