Weighted Ensemble Models Are Strong Continual Learners
Imad Eddine Marouf, Subhankar Roy, Enzo Tartaglione, and 1 more author
In European Conference on Computer Vision (ECCV) , 2024
Continual learning (CL) aims to learn from a stream of tasks without forgetting previously acquired knowledge. While most CL methods focus on mitigating catastrophic forgetting, we argue that the key to strong CL performance lies in effectively leveraging knowledge from all tasks. In this work, we propose CoFiMA, a simple yet effective approach that combines fine-tuning and model averaging. CoFiMA maintains a separate model for each task and uses a weighted ensemble for inference. We introduce a novel weighting scheme based on Fisher Information, which effectively balances the contributions of different task models. Our extensive experiments on various CL benchmarks demonstrate that CoFiMA consistently outperforms state-of-the-art methods, often by a significant margin. We provide theoretical insights into why CoFiMA works well and empirically validate its effectiveness in different CL scenarios.