DuetMerging: Synergizing Dynamic and Static Strategies for Mitigating Task Interference in Model Merging
Abstract
Model merging offers a promising paradigm for consolidating multiple expert models into a single multitask architecture. However, its effectiveness is often hindered by task interference, where conflicting parameter updates from different tasks degrade performance. While dynamic, Mixture-of-Experts based methods have improved adaptability, they are fundamentally limited by constructing their expert pools from task vectors in isolation, failing to resolve underlying structural conflicts across tasks. In this paper, we introduce DuetMerging, a novel framework that synergistically mitigates task interference from both dynamic and static perspectives. Dynamically, we apply Tucker decomposition to a unified tensor of task vectors, creating a harmonized expert pool derived from a shared core tensor that structurally enhances synergies and suppresses conflicts. Statically, we introduce a neuron-based sparsification technique that leverages task-specific neuron activation patterns to create a precise mask. This allows us to selectively preserve critical information from the decomposition's residual while suppressing functionally irrelevant or conflicting parameters. Comprehensive experiments demonstrate that DuetMerging outperforms existing methods, establishing a new state-of-the-art in both task performance and parameter efficiency.