Serviceeinschränkungen vom 12.-22.02.2026 - weitere Infos auf der UB-Homepage

Treffer: A Semantic Talking Style Space for Speech-Driven Facial Animation.

Title:
A Semantic Talking Style Space for Speech-Driven Facial Animation.
Authors:
Source:
IEEE transactions on visualization and computer graphics [IEEE Trans Vis Comput Graph] 2025 Dec; Vol. 31 (12), pp. 10801-10814.
Publication Type:
Journal Article
Language:
English
Journal Info:
Publisher: IEEE Computer Society Country of Publication: United States NLM ID: 9891704 Publication Model: Print Cited Medium: Internet ISSN: 1941-0506 (Electronic) Linking ISSN: 10772626 NLM ISO Abbreviation: IEEE Trans Vis Comput Graph Subsets: MEDLINE
Imprint Name(s):
Original Publication: New York, NY : IEEE Computer Society, c1995-
Entry Date(s):
Date Created: 20250929 Date Completed: 20251106 Latest Revision: 20251107
Update Code:
20251107
DOI:
10.1109/TVCG.2025.3615390
PMID:
41021963
Database:
MEDLINE

Weitere Informationen

We present a latent talking style space with semantic meanings for speech-driven 3D facial animation. The style space is learned from 3D speech facial animations via a self-supervision paradigm without any style labeling, leading to an automatic separation of high-level attributes, i.e., different channels of the latent style code possess different semantic meanings, such as a wide/slightly open mouth, a grinning/round mouth, and frowning/raising eyebrows. The style space enables intuitive and flexible control of talking styles in speech-driven facial animation through manipulating the channels of style code. To effectively learn such a style space, we propose a two-stage approach, involving two deep neural networks, to disentangle the person identity, speech content, and talking style contained in 3D speech facial animations. The training is performed on a novel dataset of 3D talking faces of various styles, constructed from over ten hours of videos of 200 subjects collected from the Internet.