Back to Search Start Over

Text-Guided Style Transfer-Based Image Manipulation Using Multimodal Generative Models

Authors :
Ren Togo
Megumi Kotera
Takahiro Ogawa
Miki Haseyama
Source :
IEEE Access, Vol 9, Pp 64860-64870 (2021)
Publication Year :
2021
Publisher :
IEEE, 2021.

Abstract

A new style transfer-based image manipulation framework combining generative networks and style transfer networks is presented in this paper. Unlike conventional style transfer tasks, we tackle a new task, text-guided image manipulation. We realize style transfer-based image manipulation that does not require any reference style images and generate a style image from the user’s input sentence. In our method, since an initial reference input sentence for a content image can automatically be given by an image-to-text model, the user only needs to update the reference sentence. This scheme can help users when they do not have any images representing the desired style. Although this text-guided image manipulation is a new challenging task, quantitative and qualitative comparisons showed the superiority of our method.

Details

Language :
English
ISSN :
21693536
Volume :
9
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.b7f2afac29a40e7b99c8aa28467fafa
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2021.3069876