Back to Search Start Over

SiamEFT: adaptive-time feature extraction hybrid network for RGBE multi-domain object tracking.

Authors :
Shuqi Liu
Gang Wang
Yong Song
Jinxiang Huang
Yiqian Huang
Ya Zhou
Shiqiang Wang
Source :
Frontiers in Neuroscience; 2024, p01-12, 12p
Publication Year :
2024

Abstract

Integrating RGB and Event (RGBE) multi-domain information obtained by high-dynamic-range and temporal-resolution event cameras has been considered an effective scheme for robust object tracking. However, existing RGBE tracking methods have overlooked the unique spatio-temporal features over different domains, leading to object tracking failure and ineffeciency, especally for objects against complex backgrounds. To address this problem, we propose a novel tracker based on adaptive-time feature extraction hybrid networks, namely Siamese Event Frame Tracker (SiamEFT), which focuses on the effective representation and utilization of the diverse spatio-temporal features of RGBE. We first design an adaptive-time attention module to aggregate event data into frames based on adaptive-time weights to enhance information representation. Subsequently, the SiamEF module and cross-network fusion module combining artificial neural networks and spiking neural networks hybrid network are designed to effectively extract and fuse the spatio-temporal features of RGBE. Extensive experiments on two RGBE datasets (VisEvent and COESOT) show that the SiamEFT achieves a success rate of 0.456 and 0.574, outperforming the state-of-the-art competing methods and exhibiting a 2.3-fold enhancement in effeciency. These results validate the superior accuracy and effeciency of SiamEFT in diverse and challenging scenes. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
16624548
Database :
Complementary Index
Journal :
Frontiers in Neuroscience
Publication Type :
Academic Journal
Accession number :
179161979
Full Text :
https://doi.org/10.3389/fnins.2024.1453419