Parameter-efficient Continual Learning for Enhancing Plasticity without Forgetting under Limited Model Capacity
Abstract
Avoiding catastrophic forgetting for previous tasks and maintaining model plasticity to support new tasks are two critical objectives of continual learning. However, existing methods usually neglect one of the two aspects and fail to support long task sequences with satisfactory performance, especially in resource-constrained scenarios in which the size of the model is limited. This work proposes GRAPA, a parameter-efficient continual learning method that well balances stability and plasticity of the model to handle long task sequences with diverse complexities. GRAPA enhances model plasticity without sacrificing stability with two novel designs. First, a gradient-guided parameter reuse strategy is proposed to make full use of frozen parameters while ensuring that no task interference is introduced. Second, a reinforcement-learning-based parameter allocation is designed to enable the model to adapt to the current task on top of reused parameters while preserving maximal model capacity for future tasks. Experiments on multiple task sequences composed of various datasets demonstrate that GRAPA lifts mean task accuracy by up to 7.67%, with up to 14.92% gains on subsequent complex tasks, reflecting GRAPA’s superior plasticity.