Shape-Conditioned Human Motion Diffusion Model with Mesh Representation

dc.contributor.authorXue, Kebingen_US
dc.contributor.authorSeo, Hyewonen_US
dc.contributor.authorBobenrieth, Cédricen_US
dc.contributor.authorLuo, Guoliangen_US
dc.contributor.editorBousseau, Adrienen_US
dc.contributor.editorDay, Angelaen_US
dc.date.accessioned2025-05-09T09:15:15Z
dc.date.available2025-05-09T09:15:15Z
dc.date.issued2025
dc.description.abstractHuman motion generation is a key task in computer graphics. While various conditioning signals such as text, action class, or audio have been used to harness the generation process, most existing methods neglect the case where a specific body is desired to perform the motion. Additionally, they rely on skeleton-based pose representations, necessitating additional steps to produce renderable meshes of the intended body shape. Given that human motion involves a complex interplay of bones, joints, and muscles, focusing solely on the skeleton during generation neglects the rich information carried by muscles and soft tissues, as well as their influence on movement, ultimately limiting the variability and precision of the generated motions. In this paper, we introduce Shape-conditioned Motion Diffusion model (SMD), which enables the generation of human motion directly in the form of a mesh sequence, conditioned on both a text prompt and a body mesh. To fully exploit the mesh representation while minimizing resource costs, we employ spectral representation using the graph Laplacian to encode body meshes into the learning process. Unlike retargeting methods, our model does not require source motion data and generates a variety of desired semantic motions that is inherently tailored to the given identity shape. Extensive experimental evaluations show that the SMD model not only maintains the body shape consistently with the conditioning input across motion frames but also achieves competitive performance in text-to-motion and action-to-motion tasks compared to state-of-the-art methods.en_US
dc.description.number2
dc.description.sectionheadersBringing Motion to Life: Motion Reconstruction and Control
dc.description.seriesinformationComputer Graphics Forum
dc.description.volume44
dc.identifier.doi10.1111/cgf.70065
dc.identifier.issn1467-8659
dc.identifier.pages11 pages
dc.identifier.urihttps://doi.org/10.1111/cgf.70065
dc.identifier.urihttps://diglib.eg.org/handle/10.1111/cgf70065
dc.publisherThe Eurographics Association and John Wiley & Sons Ltd.en_US
dc.subjectCCS Concepts: Computing methodologies → Modeling and simulation; Artificial intelligence
dc.subjectComputing methodologies → Modeling and simulation
dc.subjectArtificial intelligence
dc.titleShape-Conditioned Human Motion Diffusion Model with Mesh Representationen_US
Files
Original bundle
Now showing 1 - 2 of 2
Loading...
Thumbnail Image
Name:
cgf70065.pdf
Size:
1.53 MB
Format:
Adobe Portable Document Format
No Thumbnail Available
Name:
paper1217_1.mp4
Size:
37.28 MB
Format:
Video MP4