Back to Search Start Over

SAM-Net: Self-Attention based Feature Matching with Spatial Transformers and Knowledge Distillation.

Authors :
Kelenyi, Benjamin
Domsa, Victor
Tamas, Levente
Source :
Expert Systems with Applications. May2024, Vol. 242, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

In this research paper, we introduce a novel approach to enhance the performance of 2D feature matching and pose estimation through the integration of a hierarchical attention mechanism and knowledge distillation. Our proposed hierarchical attention mechanism operates at multiple scales, enabling both global context awareness and precise matching of 2D features, which is crucial for various computer vision tasks. To further improve our model's performance, we incorporate insights from an existing model PixLoc (Sarlin et al., 2021) through knowledge distillation, effectively acquiring its behavior and capabilities by ignoring dynamic objects. SAM-Net outperforms state-of-the-art methods, validated on both indoor and outdoor public datasets. For the indoor dataset, our approach achieves remarkable AUC (5 ° / 10 ° / 20 °) scores of 55.31/71.70/83.37. Similarly, for the outdoor dataset, we demonstrate outstanding AUC values of 26.01/46.44/63.61. Furthermore, SAM-Net achieves top ranking among published methods in two public visual localization benchmarks, highlighting the real benefits of the proposed method. The code and test suite can be accessed at link. 1 1 https://benjaminkelenyi.github.io/samnet/. • Geometric key-point feature extraction for 2D vision with spatial transformers. • Feature extraction with Knowledge distillation and self-attention. • Robust position estimation from multi-view camera systems. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09574174
Volume :
242
Database :
Academic Search Index
Journal :
Expert Systems with Applications
Publication Type :
Academic Journal
Accession number :
175499805
Full Text :
https://doi.org/10.1016/j.eswa.2023.122804