Back to Search Start Over

LatteGAN: Visually Guided Language Attention for Multi-Turn Text-Conditioned Image Manipulation

Authors :
Shoya Matsumori
Yuki Abe
Kosuke Shingyouchi
Komei Sugiura
Michita Imai
Source :
IEEE Access, Vol 9, Pp 160521-160532 (2021)
Publication Year :
2021
Publisher :
IEEE, 2021.

Abstract

Text-guided image manipulation tasks have recently gained attention in the vision-and-language community. While most of the prior studies focused on single-turn manipulation, our goal in this paper is to address the more challenging multi-turn image manipulation (MTIM) task. Previous models for this task successfully generate images iteratively, given a sequence of instructions and a previously generated image. However, this approach suffers from under-generation and a lack of generated quality of the objects that are described in the instructions, which consequently degrades the overall performance. To overcome these problems, we present a novel architecture called a Visually Guided Language Attention GAN (LatteGAN). Here, we address the limitations of the previous approaches by introducing a Visually Guided Language Attention (Latte) module, which extracts fine-grained text representations for the generator, and a Text-Conditioned U-Net discriminator architecture, which discriminates both the global and local representations of fake or real images. Extensive experiments on two distinct MTIM datasets, CoDraw and i-CLEVR, demonstrate the state-of-the-art performance of the proposed model. The code is available online (https://github.com/smatsumori/LatteGAN).

Details

Language :
English
ISSN :
21693536
Volume :
9
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.902b53a891e244218fca5ab31598299b
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2021.3129215