Back to Search Start Over

Fast-iTPN: Integrally Pre-Trained Transformer Pyramid Network With Token Migration.

Authors :
Tian Y
Xie L
Qiu J
Jiao J
Wang Y
Tian Q
Ye Q
Source :
IEEE transactions on pattern analysis and machine intelligence [IEEE Trans Pattern Anal Mach Intell] 2024 Dec; Vol. 46 (12), pp. 9766-9779. Date of Electronic Publication: 2024 Nov 06.
Publication Year :
2024

Abstract

We propose integrally pre-trained transformer pyramid network (iTPN), towards jointly optimizing the network backbone and the neck, so that transfer gap between representation models and downstream tasks is minimal. iTPN is born with two elaborated designs: 1) The first pre-trained feature pyramid upon vision transformer (ViT). 2) Multi-stage supervision to the feature pyramid using masked feature modeling (MFM). iTPN is updated to Fast-iTPN, reducing computational memory overhead and accelerating inference through two flexible designs. 1) Token migration: dropping redundant tokens of the backbone while replenishing them in the feature pyramid without attention operations. 2) Token gathering: reducing computation cost caused by global attention by introducing few gathering tokens. The base/large-level Fast-iTPN achieve 88.75%/89.5% top-1 accuracy on ImageNet-1 K. With 1× training schedule using DINO, the base/large-level Fast-iTPN achieves 58.4%/58.8% box AP on COCO object detection, and a 57.5%/58.7% mIoU on ADE20 K semantic segmentation using MaskDINO. Fast-iTPN can accelerate the inference procedure by up to 70%, with negligible performance loss, demonstrating the potential to be a powerful backbone for downstream vision tasks.

Details

Language :
English
ISSN :
1939-3539
Volume :
46
Issue :
12
Database :
MEDLINE
Journal :
IEEE transactions on pattern analysis and machine intelligence
Publication Type :
Academic Journal
Accession number :
39046859
Full Text :
https://doi.org/10.1109/TPAMI.2024.3429508