Back to Search Start Over

Joint estimation of pose, depth, and optical flow with a competition–cooperation transformer network.

Authors :
Liu, Xiaochen
Zhang, Tao
Liu, Mingming
Source :
Neural Networks. Mar2024, Vol. 171, p263-275. 13p.
Publication Year :
2024

Abstract

Estimating depth, ego-motion, and optical flow from consecutive frames is a critical task in robot navigation and has received significant attention in recent years. In this study, we propose PDF-Former, an unsupervised joint estimation network comprising a full transformer-based framework, as well as a competition and cooperation mechanism. The transformer framework captures global feature dependencies and is customized for different task types, thereby improving the performance of sequential tasks. The competition and cooperation mechanisms enable the network to obtain additional supervisory information at different training stages. Specifically, the competition mechanism is implemented early in training to achieve iterative optimization of 6 DOF poses (rotation and translation information from the target image to the two reference images), the depth of target image, and optical flow (from the target image to the two reference images) estimation in a competitive manner. In contrast, the cooperation mechanism is implemented later in training to facilitate the transmission of results among the three networks and mutually optimize the estimation results. We conducted experiments on the KITTI dataset, and the results indicate that PDF-Former has significant potential to enhance the accuracy and robustness of sequential tasks in robot navigation. • A transformer-based framework for joint pose, depth, and optical flow estimation is proposed. • The competition–cooperation mechanisms are proposed to promote performance improvement. • Achieved notable accuracy in public datasets. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
08936080
Volume :
171
Database :
Academic Search Index
Journal :
Neural Networks
Publication Type :
Academic Journal
Accession number :
175032182
Full Text :
https://doi.org/10.1016/j.neunet.2023.12.020