Skip to yearly menu bar Skip to main content


Poster

Learning with Dynamic Motion Blending for Versatile Motion Editing

Nan Jiang · Hongjie Li · Ziye Yuan · Zimo He · Yixin Chen · Tengyu Liu · Yixin Zhu · Siyuan Huang


Abstract:

Most text-guided motion editing methods cannot generate versatile motions as they rely on limited training triplets of original motion, edited motion, and editing instruction, which fail to cover the vast combinations of possible edits. To address this challenge, we introduce MotionCutMix, a training technique that dynamically composes a huge amount of training triplets by blending body part motions based on editing instructions. However, this technique introduces increased randomness and potential body part incoordination in the generated motions. To model such rich distribution, we propose MotionReFit, an auto-regressive diffusion model with a motion coordinator. The auto-regressive strategy reduces the window size to facilitate convergence, while the motion coordinator mitigates the artifacts of motion composition. Our model handles both spatial and temporal edits without leveraging extra motion information or LLMs. We further contribute newly captured and re-annotated datasets for multiple motion editing tasks. Experimental results demonstrate that MotionReFit excels in text-guided motion edits, closely adhering to textual directives. Furthermore, ablation studies reveal that the incorporation of MotionCutMix during training enhances the generalizability of the trained model, and does not significantly hinder training convergence.

Live content is unavailable. Log in and register to view live content