Back to Search
Start Over
An Integrated Hybrid CNN–RNN Model for Visual Description and Generation of Captions
- Source :
- Circuits, Systems, and Signal Processing. 39:776-788
- Publication Year :
- 2019
- Publisher :
- Springer Science and Business Media LLC, 2019.
-
Abstract
- Video captioning is currently considered to be one of the simplest ways to index and search data efficiently. In today’s era, suitable captioning of video images can be facilitated with deep learning architectures. The focus of past research has been on providing image captions; however, the generation of high-quality captions with suitable semantics for different scenes has not yet been achieved. Therefore, this work aims to generate well-defined and meaningful captions to images and videos by using convolutional neural networks (CNN) and recurrent neural networks in combination. Beginning with the available dataset, features of images and videos were extracted using CNN. The extracted feature vectors were then utilized to generate a language model with the involvement of long short-term memory for individual word grams. The generated meaningful captions were trained using a softmax function, for performance computation using some predefined evaluation metrics. The obtained experimental results demonstrate that the proposed model outperforms existing benchmark models.
- Subjects :
- Closed captioning
0209 industrial biotechnology
business.industry
Computer science
Applied Mathematics
Feature vector
Deep learning
ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION
Pattern recognition
02 engineering and technology
Semantics
Convolutional neural network
020901 industrial engineering & automation
Recurrent neural network
Signal Processing
Softmax function
Artificial intelligence
Language model
business
Subjects
Details
- ISSN :
- 15315878 and 0278081X
- Volume :
- 39
- Database :
- OpenAIRE
- Journal :
- Circuits, Systems, and Signal Processing
- Accession number :
- edsair.doi...........3ab73990bb41915ba1becbb39fb69d20
- Full Text :
- https://doi.org/10.1007/s00034-019-01306-8