Semantic Deep Face Models
Published in 3D International Conference on 3D Vision (3DV), 2020
Abstract
Face models built from 3D face databases are often used in computer vision and graphics tasks such as face reconstruction, replacement, tracking and manipulation. For such tasks, commonly used multi-linear morphable models, which provide semantic control over facial identity and expression, often lack quality and expressivity due to their linear nature. Deep neural networks offer the possibility of non-linear face modeling, where so far most research has focused on generating realistic facial images with less focus on 3D geometry, and methods that do produce geometry have little or no notion of semantic control, thereby limiting their artistic applicability. We present a method for nonlinear 3D face modeling using neural architectures that provides intuitive semantic control over both identity and expression by disentangling these dimensions from each other, essentially combining the benefits of both multi-linear face models and nonlinear deep face networks. The result is a powerful, semantically controllable, nonlinear, parametric face model. We demonstrate the value of our semantic deep face model with applications of 3D face synthesis, facial performance transfer, performance editing, and 2D landmark-based performance retargeting.
Bibtex:
@INPROCEEDINGS {9320344, author = {P. Chandran and D. Bradley and M. Gross and T. Beeler}, booktitle = {2020 International Conference on 3D Vision (3DV)}, title = {Semantic Deep Face Models}, year = {2020}, pages = {345-354}, doi = {10.1109/3DV50981.2020.00044}, url = {https://doi.ieeecomputersociety.org/10.1109/3DV50981.2020.00044}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, month = {nov} }