Back to Search Start Over

An Integrated Hybrid CNN–RNN Model for Visual Description and Generation of Captions

Authors :
Joel J. P. C. Rodrigues
Aditya Khamparia
Ashish Khanna
Shrasti Tiwari
Babita Pandey
Deepak Gupta
Source :
Circuits, Systems, and Signal Processing. 39:776-788
Publication Year :
2019
Publisher :
Springer Science and Business Media LLC, 2019.

Abstract

Video captioning is currently considered to be one of the simplest ways to index and search data efficiently. In today’s era, suitable captioning of video images can be facilitated with deep learning architectures. The focus of past research has been on providing image captions; however, the generation of high-quality captions with suitable semantics for different scenes has not yet been achieved. Therefore, this work aims to generate well-defined and meaningful captions to images and videos by using convolutional neural networks (CNN) and recurrent neural networks in combination. Beginning with the available dataset, features of images and videos were extracted using CNN. The extracted feature vectors were then utilized to generate a language model with the involvement of long short-term memory for individual word grams. The generated meaningful captions were trained using a softmax function, for performance computation using some predefined evaluation metrics. The obtained experimental results demonstrate that the proposed model outperforms existing benchmark models.

Details

ISSN :
15315878 and 0278081X
Volume :
39
Database :
OpenAIRE
Journal :
Circuits, Systems, and Signal Processing
Accession number :
edsair.doi...........3ab73990bb41915ba1becbb39fb69d20
Full Text :
https://doi.org/10.1007/s00034-019-01306-8