1. AvatarWild: Fully controllable head avatars in the wild
- Author
-
Shaoxu Meng, Tong Wu, Fang-Lue Zhang, Shu-Yu Chen, Yuewen Ma, Wenbo Hu, and Lin Gao
- Subjects
Neural radiance fields ,Head avatar synthesis ,Face reconstruction ,Face reenactment ,Facial animation ,Information technology ,T58.5-58.64 - Abstract
Recent advancements in the field have resulted in significant progress in achieving realistic head reconstruction and manipulation using neural radiance fields (NeRF). Despite these advances, capturing intricate facial details remains a persistent challenge. Moreover, casually captured input, involving both head poses and camera movements, introduces additional difficulties to existing methods of head avatar reconstruction. To address the challenge posed by video data captured with camera motion, we propose a novel method, AvatarWild, for reconstructing head avatars from monocular videos taken by consumer devices. Notably, our approach decouples the camera pose and head pose, allowing reconstructed avatars to be visualized with different poses and expressions from novel viewpoints. To enhance the visual quality of the reconstructed facial avatar, we introduce a view-dependent detail enhancement module designed to augment local facial details without compromising viewpoint consistency. Our method demonstrates superior performance compared to existing approaches, as evidenced by reconstruction and animation results on both multi-view and single-view datasets. Remarkably, our approach stands out by exclusively relying on video data captured by portable devices, such as smartphones. This not only underscores the practicality of our method but also extends its applicability to real-world scenarios where accessibility and ease of data capture are crucial.
- Published
- 2024
- Full Text
- View/download PDF