Back to Search Start Over

Fully Automated DCNN-Based Thermal Images Annotation Using Neural Network Pretrained on RGB Data.

Authors :
Ligocki A
Jelinek A
Zalud L
Rahtu E
Source :
Sensors (Basel, Switzerland) [Sensors (Basel)] 2021 Feb 23; Vol. 21 (4). Date of Electronic Publication: 2021 Feb 23.
Publication Year :
2021

Abstract

One of the biggest challenges of training deep neural network is the need for massive data annotation. To train the neural network for object detection, millions of annotated training images are required. However, currently, there are no large-scale thermal image datasets that could be used to train the state of the art neural networks, while voluminous RGB image datasets are available. This paper presents a method that allows to create hundreds of thousands of annotated thermal images using the RGB pre-trained object detector. A dataset created in this way can be used to train object detectors with improved performance. The main gain of this work is the novel method for fully automatic thermal image labeling. The proposed system uses the RGB camera, thermal camera, 3D LiDAR, and the pre-trained neural network that detects objects in the RGB domain. Using this setup, it is possible to run the fully automated process that annotates the thermal images and creates the automatically annotated thermal training dataset. As the result, we created a dataset containing hundreds of thousands of annotated objects. This approach allows to train deep learning models with similar performance as the common human-annotation-based methods do. This paper also proposes several improvements to fine-tune the results with minimal human intervention. Finally, the evaluation of the proposed solution shows that the method gives significantly better results than training the neural network with standard small-scale hand-annotated thermal image datasets.

Details

Language :
English
ISSN :
1424-8220
Volume :
21
Issue :
4
Database :
MEDLINE
Journal :
Sensors (Basel, Switzerland)
Publication Type :
Academic Journal
Accession number :
33672344
Full Text :
https://doi.org/10.3390/s21041552