Back to Search Start Over

High-Fidelity Neural Human Motion Transfer from Monocular Video

Authors :
Kappel, Moritz
Golyanik, Vladislav
Elgharib, Mohamed
Henningson, Jann-Ole
Seidel, Hans-Peter
Castillo, Susana
Theobalt, Christian
Magnor, Marcus
Publication Year :
2020

Abstract

Video-based human motion transfer creates video animations of humans following a source motion. Current methods show remarkable results for tightly-clad subjects. However, the lack of temporally consistent handling of plausible clothing dynamics, including fine and high-frequency details, significantly limits the attainable visual quality. We address these limitations for the first time in the literature and present a new framework which performs high-fidelity and temporally-consistent human motion transfer with natural pose-dependent non-rigid deformations, for several types of loose garments. In contrast to the previous techniques, we perform image generation in three subsequent stages, synthesizing human shape, structure, and appearance. Given a monocular RGB video of an actor, we train a stack of recurrent deep neural networks that generate these intermediate representations from 2D poses and their temporal derivatives. Splitting the difficult motion transfer problem into subtasks that are aware of the temporal motion context helps us to synthesize results with plausible dynamics and pose-dependent detail. It also allows artistic control of results by manipulation of individual framework stages. In the experimental results, we significantly outperform the state-of-the-art in terms of video realism. Our code and data will be made publicly available.<br />Comment: 14 pages, 8 figures; project page: https://graphics.tu-bs.de/publications/kappel2020high-fidelity

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2012.10974
Document Type :
Working Paper