ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10488
43
0

Streaming Generation of Co-Speech Gestures via Accelerated Rolling Diffusion

13 March 2025
Evgeniia Vu
Andrei Boiarov
Dmitry Vetrov
    VGen
ArXivPDFHTML
Abstract

Generating co-speech gestures in real time requires both temporal coherence and efficient sampling. We introduce Accelerated Rolling Diffusion, a novel framework for streaming gesture generation that extends rolling diffusion models with structured progressive noise scheduling, enabling seamless long-sequence motion synthesis while preserving realism and diversity. We further propose Rolling Diffusion Ladder Acceleration (RDLA), a new approach that restructures the noise schedule into a stepwise ladder, allowing multiple frames to be denoised simultaneously. This significantly improves sampling efficiency while maintaining motion consistency, achieving up to a 2x speedup with high visual fidelity and temporal coherence. We evaluate our approach on ZEGGS and BEAT, strong benchmarks for real-world applicability. Our framework is universally applicable to any diffusion-based gesture generation model, transforming it into a streaming approach. Applied to three state-of-the-art methods, it consistently outperforms them, demonstrating its effectiveness as a generalizable and efficient solution for real-time, high-fidelity co-speech gesture synthesis.

View on arXiv
@article{vu2025_2503.10488,
  title={ Streaming Generation of Co-Speech Gestures via Accelerated Rolling Diffusion },
  author={ Evgeniia Vu and Andrei Boiarov and Dmitry Vetrov },
  journal={arXiv preprint arXiv:2503.10488},
  year={ 2025 }
}
Comments on this paper