计算机科学
遗忘
对偶(语法数字)
人工智能
一般化
班级(哲学)
特征(语言学)
理论(学习稳定性)
渐进式学习
机器学习
人工神经网络
数学分析
艺术
哲学
文学类
语言学
数学
作者
Zhiling Fu,Zhe Wang,Xinlei Xu,Dongdong Li,Hai Yang
标识
DOI:10.1016/j.patcog.2023.109310
摘要
Most existing class incremental learning methods rely on storing old exemplars to avoid catastrophic forgetting. However, these methods inevitably face the gradient conflict problem, the inherent conflict between new streaming knowledge and existing knowledge in the gradient direction. To alleviate gradient conflict, this paper reuses the previous knowledge and expands the branch to accommodate new concepts instead of fine-tuning the original models. Specifically, this paper designs a novel dual-branch network called Knowledge Aggregation Networks. The previously trained model is frozen as a branch to retain existing knowledge, and a consistent trainable network is constructed as the other branch to learn new concepts. An adaptive feature fusion module is adopted to dynamically balance the two branches’ information during training. Moreover, a model compression stage maintains the dual-branch structure. Extensive experiments on CIFAR-100, ImageNet-Sub, and ImageNet show that our method significantly outperforms the other methods and effectively balances stability and plasticity.
科研通智能强力驱动
Strongly Powered by AbleSci AI