English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Self-supervised Multi-level Face Model Learning for Monocular Reconstruction at over 250 Hz

Tewari, A., Zollhöfer, M., Garrido, P., Bernard, F., Kim, H., Pérez, P., et al. (2017). Self-supervised Multi-level Face Model Learning for Monocular Reconstruction at over 250 Hz. Retrieved from http://arxiv.org/abs/1712.02859.

Item is

Files

show Files
hide Files
:
arXiv:1712.02859.pdf (Preprint), 4MB
Name:
arXiv:1712.02859.pdf
Description:
File downloaded from arXiv at 2018-02-02 13:46
OA-Status:
Visibility:
Public
MIME-Type / Checksum:
application/pdf / [MD5]
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

show
hide
 Creators:
Tewari, Ayush1, Author           
Zollhöfer, Michael1, Author           
Garrido, Pablo1, Author           
Bernard, Florian1, Author           
Kim, Hyeongwoo1, Author           
Pérez, Patrick2, Author
Theobalt, Christian1, Author                 
Affiliations:
1Computer Graphics, MPI for Informatics, Max Planck Society, ou_40047              
2External Organizations, ou_persistent22              

Content

show
hide
Free keywords: Computer Science, Computer Vision and Pattern Recognition, cs.CV
 Abstract: The reconstruction of dense 3D models of face geometry and appearance from a single image is highly challenging and ill-posed. To constrain the problem, many approaches rely on strong priors, such as parametric face models learned from limited 3D scan data. However, prior models restrict generalization of the true diversity in facial geometry, skin reflectance and illumination. To alleviate this problem, we present the first approach that jointly learns 1) a regressor for face shape, expression, reflectance and illumination on the basis of 2) a concurrently learned parametric face model. Our multi-level face model combines the advantage of 3D Morphable Models for regularization with the out-of-space generalization of a learned corrective space. We train end-to-end on in-the-wild images without dense annotations by fusing a convolutional encoder with a differentiable expert-designed renderer and a self-supervised training loss, both defined at multiple detail levels. Our approach compares favorably to the state-of-the-art in terms of reconstruction quality, better generalizes to real world faces, and runs at over 250 Hz.

Details

show
hide
Language(s): eng - English
 Dates: 2017-12-072017
 Publication Status: Published online
 Pages: 16 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 1712.02859
URI: http://arxiv.org/abs/1712.02859
BibTex Citekey: tewari2017
 Degree: -

Event

show

Legal Case

show

Project information

show

Source

show