1The Chinese University of Hong Kong,
2University of Washington,
3The University of British Columbia,
4UMass Amherst,
5MIT-IBM Watson AI Lab,
6Cisco Research
We introduce UniMuMo, a unified multimodal model capable of taking arbitrary text, music, and motion data as input conditions to generate outputs across all three modalities.
To address the lack of time-synchronized data, we align unpaired music and motion data based on rhythmic patterns to leverage existing large-scale music-only and motion-only datasets.
By converting music, motion, and text into token-based representation, our model bridges these modalities through a unified encoder-decoder transformer architecture.
To support multiple generation tasks within a single framework, we introduce several architectural improvements.
We propose encoding motion with a music codebook, mapping motion into the same feature space as music.
We introduce a music-motion parallel generation scheme that unifies all music and motion generation tasks into a single transformer decoder architecture with a single training task of music-motion joint generation.
Moreover, the model is designed by fine-tuning existing pre-trained single-modality models, significantly reducing computational demands.
Extensive experiments demonstrate that UniMuMo achieves competitive results on all unidirectional generation benchmarks across music, motion, and text modalities.
Methodology Overview
The training of UniMuMo consists of three stages:
In stage 1, we train a motion RVQ-VAE using the frozen codebook from a pre-trained music RVQ-VAE to encode motion into the same space as music.
In stage 2, we fine-tune a pre-trained music transformer decoder model on the text-to-music-motion task using the music-motion parallel generation scheme.
In stage 3, we fine-tune a T5 decoder for music-motion captioning using the previous music-motion decoder as a feature extractor.
More Qualitative Results
(Please turn the audio on)
Text → Music + Motion
This table presents examples of UniMuMo's text-to-music-motion generation. The text prompts are randonly selected from our synthesized music and motion descriptions and paired with each other.
Motion + Text → Music
This table presents examples of UniMuMo's motion-text-to-music generation. The motion sequences are selected from the test set of AIST++, paired randomly with our
synthesized music descriptions. Note that in the videos shown below, only the audio part is generated by ours.
Music + Text → Motion
This table presents examples of UniMuMo's music-text-to-motion generation. The music tracks are selected from the test set of Music4All, paired randomly with our
synthesized motion descriptions. Note that in the videos shown below, only the motion part is generated by ours.
Music → Text
This table presents examples of UniMuMo's music-to-text generation. The music tracks are randomly selected from the test set of MusicQA dataset.
Motion → Text
This table presents examples of UniMuMo's motion-to-text generation. The motion sequences are randomly selected
from the test set of HumanML3D dataset.
Music Motion Alignment
This table presents examples of our proposed music-motion alignment methods. In each row, a motion sequence is paired with different music tracks.
Music and motion are all from AIST++ dataset.