Poster
MC: Multi-concept Guidance for Customized Multi-concept Generation
Jiaxiu Jiang · Yabo Zhang · Kailai Feng · Xiaohe Wu · Wenbo Li · Renjing Pei · Fan Li · Wangmeng Zuo
[
Abstract
]
Abstract:
Customized text-to-image generation, which synthesizes images based on user-specified concepts, has made significant progress in handling individual concepts. However, when extended to multiple concepts, existing methods often struggle with properly integrating different models and avoiding the unintended blending of characteristics from distinct concepts. In this paper, we propose MC, a novel approach for multi-concept customization that enhances flexibility and fidelity through inference-time optimization. MC enables the integration of multiple single-concept models with heterogeneous architectures. By adaptively refining attention weights between visual and textual tokens, our method ensures that image regions accurately correspond to their associated concepts while minimizing interference between concepts. Extensive experiments demonstrate that MC outperforms training-based methods in terms of prompt-reference alignment. Furthermore, MC can be seamlessly applied to text-to-image generation, providing robust compositional capabilities. To facilitate the evaluation of multi-concept customization, we also introduce a new benchmark, MC++. The code will be publicly available.
Live content is unavailable. Log in and register to view live content