Back to Search Start Over

Learning Generative Image Manipulations from Language Instructions

Authors :
Längkvist, Martin
Persson, Andreas
Loutfi, Amy
Längkvist, Martin
Persson, Andreas
Loutfi, Amy
Publication Year :
2020

Abstract

This paper studies whether a perceptual visual system can simulate human-like cognitive capabilities by training a computational model to predict the output of an action using language instruction. The aim is to ground action words such that an AI is able to generate an output image that outputs the effect of a certain action on an given object. The output of the model is a synthetic generated image that demonstrates the effect that the action has on the scene. This work combines an image encoder, language encoder, relational network, and image generator to ground action words, and then visualize the effect an action would have on a simulated scene. The focus in this work is to learn meaningful shared image and text representations for relational learning and object manipulation.

Details

Database :
OAIster
Notes :
application/pdf, English
Publication Type :
Electronic Resource
Accession number :
edsoai.on1234158429
Document Type :
Electronic Resource