Dance Style Transfer with Cross-modal Transformer
- URL: http://arxiv.org/abs/2208.09406v3
- Date: Mon, 3 Apr 2023 08:16:37 GMT
- Title: Dance Style Transfer with Cross-modal Transformer
- Authors: Wenjie Yin, Hang Yin, Kim Baraka, Danica Kragic, and M{\aa}rten
Bj\"orkman
- Abstract summary: CycleDance is a dance style transfer system to transform an existing motion clip in one dance style to a motion clip in another dance style.
Our method extends an existing CycleGAN architecture for modeling audio sequences and integrates multimodal transformer encoders to account for music context.
- Score: 17.216186480300756
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present CycleDance, a dance style transfer system to transform an existing
motion clip in one dance style to a motion clip in another dance style while
attempting to preserve motion context of the dance. Our method extends an
existing CycleGAN architecture for modeling audio sequences and integrates
multimodal transformer encoders to account for music context. We adopt sequence
length-based curriculum learning to stabilize training. Our approach captures
rich and long-term intra-relations between motion frames, which is a common
challenge in motion transfer and synthesis work. We further introduce new
metrics for gauging transfer strength and content preservation in the context
of dance movements. We perform an extensive ablation study as well as a human
study including 30 participants with 5 or more years of dance experience. The
results demonstrate that CycleDance generates realistic movements with the
target style, significantly outperforming the baseline CycleGAN on naturalness,
transfer strength, and content preservation.
Related papers
- DanceFusion: A Spatio-Temporal Skeleton Diffusion Transformer for Audio-Driven Dance Motion Reconstruction [3.2189686611762163]
This paper introduces DanceFusion, a novel framework for reconstructing and generating dance movements synchronized to music.
The framework adeptly handles incomplete and noisy skeletal data common in short-form dance videos on social media platforms like TikTok.
Comprehensive evaluations demonstrate that DanceFusion surpasses existing methods, providing state-of-the-art performance in generating dynamic, realistic, and stylistically diverse dance motions.
arXiv Detail & Related papers (2024-11-07T12:11:11Z) - BRACE: The Breakdancing Competition Dataset for Dance Motion Synthesis [123.73677487809418]
We introduce a new dataset aiming to challenge common assumptions in dance motion synthesis.
We focus on breakdancing which features acrobatic moves and tangled postures.
Our efforts produced the BRACE dataset, which contains over 3 hours and 30 minutes of densely annotated poses.
arXiv Detail & Related papers (2022-07-20T18:03:54Z) - Bailando: 3D Dance Generation by Actor-Critic GPT with Choreographic
Memory [92.81383016482813]
We propose a novel music-to-dance framework, Bailando, for driving 3D characters to dance following a piece of music.
We introduce an actor-critic Generative Pre-trained Transformer (GPT) that composes units to a fluent dance coherent to the music.
Our proposed framework achieves state-of-the-art performance both qualitatively and quantitatively.
arXiv Detail & Related papers (2022-03-24T13:06:43Z) - Music-to-Dance Generation with Optimal Transport [48.92483627635586]
We propose a Music-to-Dance with Optimal Transport Network (MDOT-Net) for learning to generate 3D dance choreographs from music.
We introduce an optimal transport distance for evaluating the authenticity of the generated dance distribution and a Gromov-Wasserstein distance to measure the correspondence between the dance distribution and the input music.
arXiv Detail & Related papers (2021-12-03T09:37:26Z) - Rhythm is a Dancer: Music-Driven Motion Synthesis with Global Structure [47.09425316677689]
We present a music-driven motion synthesis framework that generates long-term sequences of human motions synchronized with the input beats.
Our framework enables generation of diverse motions that are controlled by the content of the music, and not only by the beat.
arXiv Detail & Related papers (2021-11-23T21:26:31Z) - Learning to dance: A graph convolutional adversarial network to generate
realistic dance motions from audio [7.612064511889756]
Learning to move naturally from music, i.e., to dance, is one of the more complex motions humans often perform effortlessly.
In this paper, we design a novel method based on graph convolutional networks to tackle the problem of automatic dance generation from audio information.
Our method uses an adversarial learning scheme conditioned on the input music audios to create natural motions preserving the key movements of different music styles.
arXiv Detail & Related papers (2020-11-25T19:53:53Z) - Learning to Generate Diverse Dance Motions with Transformer [67.43270523386185]
We introduce a complete system for dance motion synthesis.
A massive dance motion data set is created from YouTube videos.
A novel two-stream motion transformer generative model can generate motion sequences with high flexibility.
arXiv Detail & Related papers (2020-08-18T22:29:40Z) - Dance Revolution: Long-Term Dance Generation with Music via Curriculum
Learning [55.854205371307884]
We formalize the music-conditioned dance generation as a sequence-to-sequence learning problem.
We propose a novel curriculum learning strategy to alleviate error accumulation of autoregressive models in long motion sequence generation.
Our approach significantly outperforms the existing state-of-the-arts on automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-06-11T00:08:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.