Back to Search Start Over

An Image Captioning Algorithm Based on Combination Attention Mechanism.

Authors :
Liu, Jinlong
Cheng, Kangda
Jin, Haiyan
Wu, Zhilu
Source :
Electronics (2079-9292); May2022, Vol. 11 Issue 9, pN.PAG-N.PAG, 14p
Publication Year :
2022

Abstract

With the maturity of computer vision and natural language processing technology, we are becoming more ambitious in image captioning. In particular, we are more ambitious in generating longer, richer, and more accurate sentences as image descriptions. Most existing image caption models use an encoder—decoder structure, and most of the best-performing models incorporate attention mechanisms in the encoder—decoder structure. However, existing image captioning methods focus only on visual attention mechanism and not on keywords attention mechanism, thus leading to model-generated sentences that are not rich and accurate enough, and errors in visual feature extraction can directly lead to generated caption sentences that are incorrect. To fill this gap, we propose a combination attention module. This module comprises a visual attention module and a keyword attention module. The visual attention module helps in performing fast extractions of key local features, and the keyword attention module focuses on keywords that may appear in generated sentences. The results generated by the two modules can be corrected for each other. We embed the combination attention module into the framework of the Transformer, thus constructing a new image caption model CAT (Combination Attention Transformer) to generate more accurate and rich image caption sentences. Extensive experiments on the MSCOCO dataset demonstrate the effectiveness and superiority of our method over many state-of-the-art methods. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20799292
Volume :
11
Issue :
9
Database :
Complementary Index
Journal :
Electronics (2079-9292)
Publication Type :
Academic Journal
Accession number :
156849104
Full Text :
https://doi.org/10.3390/electronics11091397