Back to Search Start Over

SA‐FlowNet: Event‐based self‐attention optical flow estimation with spiking‐analogue neural networks

Authors :
Fan Yang
Li Su
Jinxiu Zhao
Xuena Chen
Xiangyu Wang
Na Jiang
Quan Hu
Source :
IET Computer Vision, Vol 17, Iss 8, Pp 925-935 (2023)
Publication Year :
2023
Publisher :
Wiley, 2023.

Abstract

Abstract Inspired by biological vision mechanism, event‐based cameras have been developed to capture continuous object motion and detect brightness changes independently and asynchronously, which overcome the limitations of traditional frame‐based cameras. Complementarily, spiking neural networks (SNNs) offer asynchronous computations and exploit the inherent sparseness of spatio‐temporal events. Notably, event‐based pixel‐wise optical flow estimations calculate the positions and relationships of objects in adjacent frames; however, as event camera outputs are sparse and uneven, dense scene information is difficult to generate and the local receptive fields of the neural network also lead to poor moving objects tracking. To address these issues, an improved event‐based self‐attention optical flow estimation network (SA‐FlowNet) that independently uses criss‐cross and temporal self‐attention mechanisms, directly capturing long‐range dependencies and efficiently extracting the temporal and spatial features from the event streams is proposed. In the former mechanism, a cross‐domain attention scheme dynamically fusing the temporal‐spatial features is introduced. The proposed network adopts a spiking‐analogue neural network architecture using an end‐to‐end learning method and gains significant computational energy benefits especially for SNNs. The state‐of‐the‐art results of the error rate for optical flow prediction on the Multi‐Vehicle Stereo Event Camera (MVSEC) dataset compared with the current SNN‐based approaches is demonstrated.

Details

Language :
English
ISSN :
17519640 and 17519632
Volume :
17
Issue :
8
Database :
Directory of Open Access Journals
Journal :
IET Computer Vision
Publication Type :
Academic Journal
Accession number :
edsdoj.05714d28e2f74398a43eaba8c904f0e0
Document Type :
article
Full Text :
https://doi.org/10.1049/cvi2.12206