Back to Search Start Over

Vital information matching in vision-and-language navigation.

Authors :
Jia Z
Yu K
Ru J
Yang S
Coleman S
Source :
Frontiers in neurorobotics [Front Neurorobot] 2022 Nov 17; Vol. 16, pp. 1035921. Date of Electronic Publication: 2022 Nov 17 (Print Publication: 2022).
Publication Year :
2022

Abstract

With the rapid development of artificial intelligence technology, many researchers have begun to focus on visual language navigation, which is one of the most important tasks in multi-modal machine learning. The focus of this multi-modal field is how to fuse multiple inputs, which is crucial for the integrated feedback of intrinsic information. However, the existing models are only implemented through simple data augmentation or expansion, and are obviously far from being able to tap the intrinsic relationship between modalities. In this paper, to overcome these challenges, a novel multi-modal matching feedback self-tuning model is proposed, which is a novel neural network called Vital Information Matching Feedback Self-tuning Network (VIM-Net). Our VIM-Net network is mainly composed of two matching feedback modules, a visual matching feedback module (V-mat) and a trajectory matching feedback module (T-mat). Specifically, V-mat matches the target information of visual recognition with the entity information extracted by the command; T-mat matches the serialized trajectory feature with the direction of movement of the command. Ablation experiments and comparative experiments are conducted on the proposed model using the Matterport3D simulator and the Room-to-Room (R2R) benchmark datasets, and the final navigation effect is shown in detail. The results prove that the model proposed in this paper is indeed effective on the task.<br />Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.<br /> (Copyright © 2022 Jia, Yu, Ru, Yang and Coleman.)

Details

Language :
English
ISSN :
1662-5218
Volume :
16
Database :
MEDLINE
Journal :
Frontiers in neurorobotics
Publication Type :
Academic Journal
Accession number :
36467568
Full Text :
https://doi.org/10.3389/fnbot.2022.1035921