Back to Search Start Over

Leveraging Diverse Vectors in ViT for Image Super Resolution.

Authors :
Shijie Liu
Wenchong Wu
Yungang Zhang
Source :
International Journal of Design, Analysis & Tools for Integrated Circuits & Systems; Dec2023, Vol. 12 Issue 2, p30-36, 7p
Publication Year :
2023

Abstract

Vector serves as the fundamental element in the Attention calculation of the Vision Transformer (ViT). In the conventional ViT approach, spatial points within the feature map are treated as vectors, with each spatial point representing a point in h × w dimensions. We aim to broaden the spectrum of vectors and explore various forms, such as considering numerical values of feature map, channels, or windows as vectors for Attention calculation. We propose a hybrid ViT model named the High to Low Frequency Attention Network (HLFANet). Through Single Image Super-Resolution (SISR) experiments, it has been demonstrated that the different forms of vectors complement each other in learning and representing high and low-frequency features. Additionally, utilizing mixed attention helps mitigate the trade-off between computational cost and performance associated with window-based attention. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20712987
Volume :
12
Issue :
2
Database :
Complementary Index
Journal :
International Journal of Design, Analysis & Tools for Integrated Circuits & Systems
Publication Type :
Academic Journal
Accession number :
176955423