Skip to yearly menu bar Skip to main content


Poster

ConMo: Controllable Motion Disentanglement and Recomposition for Zero-Shot Motion Transfer

Jiayi Gao · Zijin Yin · Changcheng Hua · Yuxin Peng · Kongming Liang · Zhanyu Ma · Jun Guo · Yang Liu


Abstract:

The development of Text-to-Video (T2V) generation has made motion transfer possible, enabling the control of video motion based on existing footage. However, current techniques struggle to preserve the diversity and accuracy of motion while transferring it to subjects with varying shapes. Additionally, they fail to handle videos with multiple subjects, making it difficult to disentangle and control motion transfer to specific instances.To overcome these challenges, we introduce ConMo, a training-free framework that disentangle and recompose the motions of subjects and camera movements. Unlike existing methods, ConMo separates individual subject motions and background motions from compound motion trajectories in complex videos, allowing for precise control and flexible manipulation, such as resizing and repositioning. This leads to more accurate motion transfer across subjects of different shapes and better handling of videos with multiple subjects. Extensive experiments demonstrate that ConMo can effectively achieve the aforementioned various applications and surpass existing state-of-the-art methods in terms of motion fidelity and temporal consistency. The code will be published.

Live content is unavailable. Log in and register to view live content