1. LDLS: 3-D Object Segmentation Through Label Diffusion From 2-D Images
- Author
-
Bharath Hariharan, Wei-Lun Chao, Brian H. Wang, Yan Wang, Mark Campbell, and Kilian Q. Weinberger
- Subjects
FOS: Computer and information sciences ,Control and Optimization ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Biomedical Engineering ,Point cloud ,02 engineering and technology ,Computer Science - Robotics ,Artificial Intelligence ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Segmentation ,Computer vision ,Pixel ,business.industry ,Mechanical Engineering ,Deep learning ,Image and Video Processing (eess.IV) ,020206 networking & telecommunications ,Mobile robot ,Image segmentation ,Electrical Engineering and Systems Science - Image and Video Processing ,Object (computer science) ,Computer Science Applications ,Human-Computer Interaction ,Control and Systems Engineering ,Graph (abstract data type) ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Robotics (cs.RO) - Abstract
Object segmentation in three-dimensional (3-D) point clouds is a critical task for robots capable of 3-D perception. Despite the impressive performance of deep learning-based approaches on object segmentation in 2-D images, deep learning has not been applied nearly as successfully for 3-D point cloud segmentation. Deep networks generally require large amounts of labeled training data, which are readily available for 2-D images but are difficult to produce for 3-D point clouds. In this letter, we present Label Diffusion Lidar Segmentation (LDLS), a novel approach for 3-D point cloud segmentation, which leverages 2-D segmentation of an RGB image from an aligned camera to avoid the need for training on annotated 3-D data. We obtain 2-D segmentation predictions by applying Mask-RCNN to the RGB image, and then link this image to a 3-D lidar point cloud by building a graph of connections among 3-D points and 2-D pixels. This graph then directs a semi-supervised label diffusion process, where the 2-D pixels act as source nodes that diffuse object label information through the 3-D point cloud, resulting in a complete 3-D point cloud segmentation. We conduct empirical studies on the KITTI benchmark dataset and on a mobile robot, demonstrating wide applicability and superior performance of LDLS compared with the previous state of the art in 3-D point cloud segmentation, without any need for either 3-D training data or fine tuning of the 2-D image segmentation model., Comment: Accepted for publication in IEEE Robotics and Automation Letters with presentation at IROS 2019
- Published
- 2019
- Full Text
- View/download PDF