计算机研究与发展2024,Vol.61Issue(6) :1476-1496.DOI:10.7544/issn1000-1239.202220820

持续学习的研究进展与趋势

Advances and Trends of Continual Learning

李文斌 熊亚锟 范祉辰 邓波 曹付元 高阳
计算机研究与发展2024,Vol.61Issue(6) :1476-1496.DOI:10.7544/issn1000-1239.202220820

持续学习的研究进展与趋势

Advances and Trends of Continual Learning

李文斌 1熊亚锟 1范祉辰 1邓波 2曹付元 3高阳1
扫码查看

作者信息

  • 1. 计算机软件新技术国家重点实验室(南京大学) 南京 210023
  • 2. 中国科学院空天信息创新研究院 北京 100094
  • 3. 山西大学计算机与信息技术学院(大数据学院) 太原 030006;山西大学大数据学院 太原 030006
  • 折叠

摘要

随着深度学习技术的发展与应用,特别是资源受限场景和数据安全场景对序列任务和数据进行快速学习需求的增多,持续学习逐渐成为机器学习领域关注的一个新热点.不同于人类所具备的持续学习和迁移知识的能力,现有深度学习模型在序列学习过程中容易遭受灾难性遗忘的问题.因此,如何在动态、非平稳的序列任务及流式数据中不断学习新知识、同时保留旧知识是持续学习研究的核心.首先,通过对近年来持续学习国内外相关工作的调研与总结,将持续学习方法分为基于回放、基于约束、基于结构三大类,并对这 3类方法做进一步的细分.具体而言,根据所使用的样本来源将基于回放的方法细分为采样回放、生成回放、伪样本回放 3类;根据训练约束的来源将基于约束的方法细分为参数约束、梯度约束、数据约束 3类;根据对于模型结构的使用方式将基于结构的方法细分为参数隔离、模型拓展 2类.通过对比相关工作的创新点,对各类方法的优缺点进行总结.其次,对国内外研究现状进行分析.最后,针对持续学习与其他领域相结合的未来发展方向进行展望.

Abstract

With the development and successful application of deep learning,continual learning has attracted increasing attention and has been a hot topic in the field of machine learning,especially in the resource-limited and data-security scenarios with the increasing requirements of quickly learning sequential tasks and data.Unlike humans who enjoy the ability of continually learning and transferring knowledge,the existing deep learning models are prone to easily suffering from a catastrophic forgetting problem in a sequential learning process.Therefore,how to continually learn new knowledge and retain old knowledge at the same time on dynamic and non-stationary sequential task and streaming data,is the core of continual learning.Firstly,through the investigation and summary of the related work of continual learning at home and abroad in recent years,continual learning methods can be roughly divided into three categories:replay-based,constraint-based,and architecture-based.We further subdivide these three types of methods.Specifically,the replay-based methods are subdivided into three categories:sample replay,generation replay,and pseudo-sample replay,according to the sample's sources used;the constraint-based methods are subdivided into parameter constraints,gradient constraints,and data constraints,according to the constraint's sources;the architecture-based methods are subdivided into two categories:parameter isolation and model expansion,according to how the model structure is used.By comparing the innovation points of related work,the advantages and disadvantages of various methods are summarized.Secondly,the research progress at home and abroad is analyzed.Finally,the future development direction of continual learning combined with other fields is simply prospected.

关键词

深度学习/知识迁移/持续学习/灾难性遗忘/序列任务

Key words

deep learning/knowledge transfer/continual learning/catastrophic forgetting/sequential task

引用本文复制引用

基金项目

国家自然科学基金(62192783)

国家自然科学基金(62106100)

江苏省自然科学基金(BK20221441)

江苏省双创博士计划(JSSCBS20210021)

出版年

2024
计算机研究与发展
中国科学院计算技术研究所 中国计算机学会

计算机研究与发展

CSTPCDCSCD北大核心
影响因子:2.649
ISSN:1000-1239
被引量1
参考文献量94
段落导航相关论文