Finger dance is an emerging social media trend using finger gesture motions for expression. Music to finger dance generation is challenging due to its fine-grained movements. Existing music-driven methods often fail to model subtle finger motions, yielding poor performances. We propose Fine-Finger Diffusion (FFD), the first end-to-end framework for music to finger dance generation. Our method employs a diffusion model to create rhythmically aligned finger movements while ensuring motion stability. A novel detail-aware loss (DAL) enhances temporal coherence by constraining inter-frame motion fluctuations. We introduce DanceFingers-4K, the first large-scale finger dance dataset containing 4007 video clips with music-motion pairs. Comprehensive evaluations demonstrate FFD’s superiority over existing approaches across objective metrics and user study.
FFD: Fine-Finger Diffusion Model for Music to Fine-grained Finger Dance Generation
Boyan Dong,Wen-Ling Lei,Li Liu
Published 2025 in Interspeech
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
Interspeech
- Publication date
2025-08-17
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-28 of 28 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1