Facial Animation with Disentangled Identity and Motion using Transformers
dc.contributor.author | Chandran, Prashanth | en_US |
dc.contributor.author | Zoss, Gaspard | en_US |
dc.contributor.author | Gross, Markus | en_US |
dc.contributor.author | Gotardo, Paulo | en_US |
dc.contributor.author | Bradley, Derek | en_US |
dc.contributor.editor | Dominik L. Michels | en_US |
dc.contributor.editor | Soeren Pirk | en_US |
dc.date.accessioned | 2022-08-10T15:19:57Z | |
dc.date.available | 2022-08-10T15:19:57Z | |
dc.date.issued | 2022 | |
dc.description.abstract | We propose a 3D+time framework for modeling dynamic sequences of 3D facial shapes, representing realistic non-rigid motion during a performance. Our work extends neural 3D morphable models by learning a motion manifold using a transformer architecture. More specifically, we derive a novel transformer-based autoencoder that can model and synthesize 3D geometry sequences of arbitrary length. This transformer naturally determines frame-to-frame correlations required to represent the motion manifold, via the internal self-attention mechanism. Furthermore, our method disentangles the constant facial identity from the time-varying facial expressions in a performance, using two separate codes to represent neutral identity and the performance itself within separate latent subspaces. Thus, the model represents identity-agnostic performances that can be paired with an arbitrary new identity code and fed through our new identity-modulated performance decoder; the result is a sequence of 3D meshes for the performance with the desired identity and temporal length. We demonstrate how our disentangled motion model has natural applications in performance synthesis, performance retargeting, key-frame interpolation and completion of missing data, performance denoising and retiming, and other potential applications that include full 3D body modeling. | en_US |
dc.description.number | 8 | |
dc.description.sectionheaders | Capture, Tracking, and Facial Animation | |
dc.description.seriesinformation | Computer Graphics Forum | |
dc.description.volume | 41 | |
dc.identifier.doi | 10.1111/cgf.14641 | |
dc.identifier.issn | 1467-8659 | |
dc.identifier.pages | 267-277 | |
dc.identifier.pages | 11 pages | |
dc.identifier.uri | https://doi.org/10.1111/cgf.14641 | |
dc.identifier.uri | https://diglib.eg.org:443/handle/10.1111/cgf14641 | |
dc.publisher | The Eurographics Association and John Wiley & Sons Ltd. | en_US |
dc.subject | CCS Concepts: Computing methodologies --> Motion processing; Shape modeling; Mesh geometry models | |
dc.subject | Computing methodologies | |
dc.subject | Motion processing | |
dc.subject | Shape modeling | |
dc.subject | Mesh geometry models | |
dc.title | Facial Animation with Disentangled Identity and Motion using Transformers | en_US |
Files
Original bundle
1 - 1 of 1