ZeroEGGS: Zero‐shot Example‐based Gesture Generation from Speech
dc.contributor.author | Ghorbani, Saeed | en_US |
dc.contributor.author | Ferstl, Ylva | en_US |
dc.contributor.author | Holden, Daniel | en_US |
dc.contributor.author | Troje, Nikolaus F. | en_US |
dc.contributor.author | Carbonneau, Marc‐André | en_US |
dc.contributor.editor | Hauser, Helwig and Alliez, Pierre | en_US |
dc.date.accessioned | 2023-03-22T15:07:14Z | |
dc.date.available | 2023-03-22T15:07:14Z | |
dc.date.issued | 2023 | |
dc.description.abstract | We present ZeroEGGS, a neural network framework for speech‐driven gesture generation with zero‐shot style control by example. This means style can be controlled via only a short example motion clip, even for motion styles unseen during training. Our model uses a Variational framework to learn a style embedding, making it easy to modify style through latent space manipulation or blending and scaling of style embeddings. The probabilistic nature of our framework further enables the generation of a variety of outputs given the input, addressing the stochastic nature of gesture motion. In a series of experiments, we first demonstrate the flexibility and generalizability of our model to new speakers and styles. In a user study, we then show that our model outperforms previous state‐of‐the‐art techniques in naturalness of motion, appropriateness for speech, and style portrayal. Finally, we release a high‐quality dataset of full‐body gesture motion including fingers, with speech, spanning across 19 different styles. Our code and data are publicly available at . | en_US |
dc.description.number | 1 | |
dc.description.sectionheaders | Articles | |
dc.description.seriesinformation | Computer Graphics Forum | |
dc.description.volume | 42 | |
dc.identifier.doi | 10.1111/cgf.14734 | |
dc.identifier.issn | 1467-8659 | |
dc.identifier.pages | 206-216 | |
dc.identifier.uri | https://doi.org/10.1111/cgf.14734 | |
dc.identifier.uri | https://diglib.eg.org:443/handle/10.1111/cgf14734 | |
dc.publisher | Eurographics ‐ The European Association for Computer Graphics and John Wiley & Sons Ltd. | en_US |
dc.rights | CC BY-NC-ND Attribution-NonCommercial-NoDerivatives 4.0 International | |
dc.rights.uri | https://creativecommons.org/licenses/by-nc-nd/4.0/ | |
dc.subject | animation | |
dc.subject | gestures | |
dc.subject | character control | |
dc.subject | motion capture | |
dc.title | ZeroEGGS: Zero‐shot Example‐based Gesture Generation from Speech | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- v42i1pp206-216-cgf14734.pdf
- Size:
- 1.42 MB
- Format:
- Adobe Portable Document Format