Representing Animatable Avatar via Factorized Neural Fields

dc.contributor.authorSong, Chunjinen_US
dc.contributor.authorWu, Zhijieen_US
dc.contributor.authorWandt, Bastianen_US
dc.contributor.authorSigal, Leoniden_US
dc.contributor.authorRhodin, Helgeen_US
dc.contributor.editorAttene, Marcoen_US
dc.contributor.editorSellán, Silviaen_US
dc.date.accessioned2025-06-20T07:40:21Z
dc.date.available2025-06-20T07:40:21Z
dc.date.issued2025
dc.description.abstractFor reconstructing high-fidelity human 3D models from monocular videos, it is crucial to maintain consistent large-scale body shapes along with finely matched subtle wrinkles. This paper explores how per-frame rendering results can be factorized into a pose-independent component and a corresponding pose-dependent counterpart to facilitate frame consistency at multiple scales. Pose adaptive texture features are further improved by restricting the frequency bands of these two components. Pose-independent outputs are expected to be low-frequency, while high-frequency information is linked to pose-dependent factors. We implement this with a dual-branch network. The first branch takes coordinates in the canonical space as input, while the second one additionally considers features outputted by the first branch and pose information of each frame. A final network integrates the information predicted by both branches and utilizes volume rendering to generate photo-realistic 3D human images. Through experiments, we demonstrate that our method consistently surpasses all state-of-the-art methods in preserving high-frequency details and ensuring consistent body contours. Our code is accessible at https://github.com/ChunjinSong/facavatar.en_US
dc.description.number5
dc.description.sectionheadersAnimation and Morphing
dc.description.seriesinformationComputer Graphics Forum
dc.description.volume44
dc.identifier.doi10.1111/cgf.70192
dc.identifier.issn1467-8659
dc.identifier.pages13 pages
dc.identifier.urihttps://doi.org/10.1111/cgf.70192
dc.identifier.urihttps://diglib.eg.org/handle/10.1111/cgf70192
dc.publisherThe Eurographics Association and John Wiley & Sons Ltd.en_US
dc.rightsAttribution 4.0 International License
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectCCS Concepts: Computing methodologies → Reconstruction; Shape inference
dc.subjectComputing methodologies → Reconstruction
dc.subjectShape inference
dc.titleRepresenting Animatable Avatar via Factorized Neural Fieldsen_US
Files
Original bundle
Now showing 1 - 2 of 2
Loading...
Thumbnail Image
Name:
cgf70192.pdf
Size:
3.21 MB
Format:
Adobe Portable Document Format
No Thumbnail Available
Name:
fp1-1011_mm.zip
Size:
68.5 MB
Format:
Zip file
Collections