Perceiving Systems Conference Paper 2022

I M Avatar: Implicit Morphable Head Avatars from Videos

Imavatarteaser

Traditional 3D morphable face models (3DMMs) provide fine-grained control over expression but cannot easily capture geometric and appearance details. Neural volumetric representations approach photorealism but are hard to animate and do not generalize well to unseen expressions. To tackle this problem, we propose IMavatar (Implicit Morphable avatar), a novel method for learning implicit head avatars from monocular videos. Inspired by the fine-grained control mechanisms afforded by conventional 3DMMs, we represent the expression- and pose- related deformations via learned blendshapes and skinning fields. These attributes are pose-independent and can be used to morph the canonical geometry and texture fields given novel expression and pose parameters. We employ ray marching and iterative root-finding to locate the canonical surface intersection for each pixel. A key contribution is our novel analytical gradient formulation that enables end-to-end training of IMavatars from videos. We show quantitatively and qualitatively that our method improves geometry and covers a more complete expression space compared to state-of-the-art methods.

Author(s): Zheng, Yufeng and Fernández Abrevaya, Victoria and Bühler, Marcel C and Chen, Xu and Black, Michael J and Hilliges, Otmar
Book Title: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022)
Pages: 13535--13545
Year: 2022
Month: June
Publisher: IEEE
Bibtex Type: Conference Paper (inproceedings)
Address: Piscataway, NJ
DOI: 10.1109/CVPR52688.2022.01318
Event Name: IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022)
Event Place: New Orleans, Louisiana
State: Published
URL: https://ait.ethz.ch/imavatar
Electronic Archiving: grant_archive
ISBN: 978-1-6654-6947-0
Links:

BibTex

@inproceedings{Zheng:CVPR:2022,
  title = {{I M} Avatar: Implicit Morphable Head Avatars from Videos},
  booktitle = {2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022)},
  abstract = {Traditional 3D morphable face models (3DMMs) provide fine-grained control over expression but cannot easily capture geometric and appearance details. Neural volumetric representations approach photorealism but are hard to animate and do not generalize well to unseen expressions. To tackle this problem, we propose IMavatar (Implicit Morphable avatar), a novel method for learning implicit head avatars from monocular videos. Inspired by the fine-grained control mechanisms afforded by conventional 3DMMs, we represent the expression- and pose- related deformations via learned blendshapes and skinning fields. These attributes are pose-independent and can be used to morph the canonical geometry and texture fields given novel expression and pose parameters. We employ ray marching and iterative root-finding to locate the canonical surface intersection for each pixel. A key contribution is our novel analytical gradient formulation that enables end-to-end training of IMavatars from videos. We show quantitatively and qualitatively that our method improves geometry and covers a more complete expression space compared to state-of-the-art methods.},
  pages = {13535--13545},
  publisher = {IEEE},
  address = {Piscataway, NJ},
  month = jun,
  year = {2022},
  slug = {zheng-cvpr-2022},
  author = {Zheng, Yufeng and Fern{\'a}ndez Abrevaya, Victoria and B{\"u}hler, Marcel C and Chen, Xu and Black, Michael J and Hilliges, Otmar},
  url = {https://ait.ethz.ch/imavatar},
  month_numeric = {6}
}