Back to Search
Start Over
LiT: Zero-Shot Transfer with Locked-image text Tuning
- Source :
- 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
- Publication Year :
- 2022
- Publisher :
- IEEE, 2022.
-
Abstract
- This paper presents contrastive-tuning, a simple method employing contrastive training to align image and text models while still taking advantage of their pre-training. In our empirical study we find that locked pre-trained image models with unlocked text models work best. We call this instance of contrastive-tuning "Locked-image Tuning" (LiT), which just teaches a text model to read out good representations from a pre-trained image model for new tasks. A LiT model gains the capability of zero-shot transfer to new vision tasks, such as image classification or retrieval. The proposed LiT is widely applicable; it works reliably with multiple pre-training methods (supervised and unsupervised) and across diverse architectures (ResNet, Vision Transformers and MLP-Mixer) using three different image-text datasets. With the transformer-based pre-trained ViT-g/14 model, the LiT model achieves 85.2% zero-shot transfer accuracy on the ImageNet test set, and 82.5% on the challenging out-of-distribution ObjectNet test set.<br />Comment: Xiaohua, Xiao, Basil, Andreas and Lucas contributed equally; CVPR 2022
- Subjects :
- FOS: Computer and information sciences
Computer Science - Machine Learning
Computer Science - Computation and Language
Computer Vision and Pattern Recognition (cs.CV)
Computer Science - Computer Vision and Pattern Recognition
ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION
Computation and Language (cs.CL)
Machine Learning (cs.LG)
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
- Accession number :
- edsair.doi.dedup.....eb699ad0feae4f780051c9be69643b66
- Full Text :
- https://doi.org/10.1109/cvpr52688.2022.01759