Geometry-Guided Dense Perspective Network for Speech-Driven Facial Animation.
Journal
IEEE transactions on visualization and computer graphics
ISSN: 1941-0506
Titre abrégé: IEEE Trans Vis Comput Graph
Pays: United States
ID NLM: 9891704
Informations de publication
Date de publication:
12 2022
12 2022
Historique:
pubmed:
28
8
2021
medline:
29
10
2022
entrez:
27
8
2021
Statut:
ppublish
Résumé
Realistic speech-driven 3D facial animation is a challenging problem due to the complex relationship between speech and face. In this paper, we propose a deep architecture, called Geometry-guided Dense Perspective Network (GDPnet), to achieve speaker-independent realistic 3D facial animation. The encoder is designed with dense connections to strengthen feature propagation and encourage the re-use of audio features, and the decoder is integrated with an attention mechanism to adaptively recalibrate point-wise feature responses by explicitly modeling interdependencies between different neuron units. We also introduce a non-linear face reconstruction representation as a guidance of latent space to obtain more accurate deformation, which helps solve the geometry-related deformation and is good for generalization across subjects. Huber and HSIC (Hilbert-Schmidt Independence Criterion) constraints are adopted to promote the robustness of our model and to better exploit the non-linear and high-order correlations. Experimental results on the public dataset and real scanned dataset validate the superiority of our proposed GDPnet compared with state-of-the-art model. The code is available for research purposes at http://cic.tju.edu.cn/faculty/likun/projects/GDPnet.
Identifiants
pubmed: 34449390
doi: 10.1109/TVCG.2021.3107669
doi:
Types de publication
Journal Article
Research Support, Non-U.S. Gov't
Langues
eng
Sous-ensembles de citation
IM