计算机科学
人工智能
概化理论
机器学习
序列学习
遗忘
多任务学习
任务(项目管理)
透明度(行为)
人工神经网络
基于实例的学习
主动学习
竞争性学习
在线机器学习
主动学习(机器学习)
机器人学习
机器人
统计
语言学
哲学
计算机安全
经济
管理
数学
移动机器人
作者
Mahsa Paknezhad,Hamsawardhini Rengarajan,Chenghao Yuan,Sujanya Suresh,Manas Gupta,Savitha Ramasamy,Lee Hwee Kuan
标识
DOI:10.1016/j.neunet.2023.02.007
摘要
This paper takes a parallel learning approach in continual learning scenarios. We define parallel continual learning as learning a sequence of tasks where the data for the previous tasks, whose distribution may have shifted over time, are also available while learning new tasks. We propose a parallel continual learning method by assigning subnetworks to each task, and simultaneously training only the assigned subnetworks on their corresponding tasks. In doing so, some parts of the network will be shared across multiple tasks. This is unlike the existing literature in continual learning which aims at learning incoming tasks sequentially, with the assumption that the data for the previous tasks have a fixed distribution. Our proposed method offers promises in: (1) Transparency in the network and in the relationship across tasks by enabling examination of the learned representations by independent and shared subnetworks, (2) Representation generalizability through sharing and training subnetworks on multiple tasks simultaneously. Our analysis shows that compared to many competing approaches such as continual learning, neural architecture search, and multi-task learning, parallel continual learning is capable of learning more generalizable representations. Also, (3)Parallel continual learning overcomes the common issue of catastrophic forgetting in continual learning algorithms. This is the first effort to train a neural network on multiple tasks and input domains simultaneously in a continual learning scenario. Our code is available at https://github.com/yours-anonym/PaRT.
科研通智能强力驱动
Strongly Powered by AbleSci AI