EmoDiffGes: Emotion-Aware Co-Speech Holistic Gesture Generation with Progressive Synergistic Diffusion

dc.contributor.authorLi, Xinruen_US
dc.contributor.authorLin, Jingzhongen_US
dc.contributor.authorZhang, Bohaoen_US
dc.contributor.authorQi, Yuanyuanen_US
dc.contributor.authorWang, Changboen_US
dc.contributor.authorHe, Gaoqien_US
dc.contributor.editorChristie, Marcen_US
dc.contributor.editorPietroni, Nicoen_US
dc.contributor.editorWang, Yu-Shuenen_US
dc.date.accessioned2025-10-07T05:03:04Z
dc.date.available2025-10-07T05:03:04Z
dc.date.issued2025
dc.description.abstractCo-speech gesture generation, driven by emotional expression and synergistic bodily movements, is essential for applications such as virtual avatars and human-robot interaction. Existing co-speech gesture generation methods face two fundamental limitations: (1) producing inexpressive gestures due to ignoring the temporal evolution of emotion; (2) generating incoherent and unnatural motions as a result of either holistic body oversimplification or independent part modeling. To address the above limitations, we propose EmoDiffGes, a diffusion-based framework grounded in embodied emotion theory, unifying dynamic emotion conditioning and part-aware synergistic modeling. Specifically, a Dynamic Emotion-Alignment Module (DEAM) is first applied to extract dynamic emotional cues and inject emotion guidance into the generation process. Then, a Progressive Synergistic Gesture Generator (PSGG) iteratively refines region-specific latent codes while maintaining full-body coordination, leveraging a Body Region Prior for part-specific encoding and Progressive Inter-Region Synergistic Flow for global motion coherence. Extensive experiments validate the effectiveness of our methods, showcasing the potential for generating expressive, coordinated, and emotionally grounded human gestures.en_US
dc.description.number7
dc.description.sectionheadersDigital Human
dc.description.seriesinformationComputer Graphics Forum
dc.description.volume44
dc.identifier.doi10.1111/cgf.70261
dc.identifier.issn1467-8659
dc.identifier.pages13 pages
dc.identifier.urihttps://doi.org/10.1111/cgf.70261
dc.identifier.urihttps://diglib.eg.org/handle/10.1111/cgf70261
dc.publisherThe Eurographics Association and John Wiley & Sons Ltd.en_US
dc.subjectCCS Concepts: Computing methodologies → Computer graphics; Animation; Motion processing
dc.subjectComputing methodologies → Computer graphics
dc.subjectAnimation
dc.subjectMotion processing
dc.titleEmoDiffGes: Emotion-Aware Co-Speech Holistic Gesture Generation with Progressive Synergistic Diffusionen_US
Files
Original bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
cgf70261.pdf
Size:
13.57 MB
Format:
Adobe Portable Document Format
No Thumbnail Available
Name:
paper1396_mm1.zip
Size:
332.8 MB
Format:
Zip file
Collections