Back to Search Start Over

Robust multiple-sensing-modality data fusion using Gaussian Process Implicit Surfaces

Authors :
Gerardo-Castro, MP
Peynot, T
Ramos, F
Fitch, R
Gerardo-Castro, MP
Peynot, T
Ramos, F
Fitch, R
Publication Year :
2014

Abstract

© 2014 International Society of Information Fusion. The ability to build high-fidelity 3D representations of the environment from sensor data is critical for autonomous robots. Multi-sensor data fusion allows for more complete and accurate representations. Furthermore, using distinct sensing modalities (i.e. sensors using a different physical process and/or operating at different electromagnetic frequencies) usually leads to more reliable perception, especially in challenging environments, as modalities may complement each other. However, they may react differently to certain materials or environmental conditions, leading to catastrophic fusion. In this paper, we propose a new method to reliably fuse data from multiple sensing modalities, including in situations where they detect different targets. We first compute distinct continuous surface representations for each sensing modality, with uncertainty, using Gaussian Process Implicit Surfaces (GPIS). Second, we perform a local consistency test between these representations, to separate consistent data (i.e. data corresponding to the detection of the same target by the sensors) from inconsistent data. The consistent data can then be fused together, using another GPIS process, and the rest of the data can be combined as appropriate. The approach is first validated using synthetic data. We then demonstrate its benefit using a mobile robot, equipped with a laser scanner and a radar, which operates in an outdoor environment in the presence of large clouds of airborne dust and smoke.

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1197437673
Document Type :
Electronic Resource