首页|资源受限的大模型高效迁移学习算法研究综述

资源受限的大模型高效迁移学习算法研究综述

扫码查看
近年来,深度学习在自然语言理解、计算机视觉和数据挖掘等重要领域取得了巨大成功,极大地推动了人工智能技术的发展.迁移学习的诞生和应用更是大幅减轻了数据的获取和标注成本,成倍提升了深度模型和算法的泛化能力和适用性.然而,随着模型规模的不断增大,传统的迁移学习方法面临着计算和存储资源的巨大挑战,难以满足可穿戴、军事、医疗等资源受限场景下的应用需求.高效迁移学习算法应运而生,旨在以最小的资源开销实现大模型的快速适配与部署,有望成为未来人工智能技术发展的关键突破口.本文是高效迁移学习领域的首篇中文综述,系统总结了近5年来该领域的研究进展.本文首先分析了高效迁移学习算法在自然语言处理、计算机视觉和多模态模型三大场景下的应用现状,提炼出了修改模型结构、调整预训练参数、调整原始输入(输出)、注入自适应参数、引入自适应模块等五类具有代表性的技术路线.在此基础上,本文对各类方法进行了全面梳理与比较,分析了它们的优势与局限性.本文的主要贡献如下:(1)对高效迁移学习领域进行了系统化的综述,为后续研究提供了完整的技术参考;(2)提出了一种基于技术路线的分类框架,帮助读者快速把握该领域的研究脉络;(3)深入分析了现有方法的不足,并展望了未来的发展方向,具有一定的前瞻性和指导意义.高效迁移学习算法是推动现代人工智能技术走进千家万户的关键技术,有望让更多中小企业和个人用户受益于大模型的强大性能.本文对该领域的全面梳理,将为该领域算法的进一步发展和应用提供重要的理论参考与实践指导.
Efficient Transfer Learning of Large Models with Limited Resources:A Survey
In recent years,the fast-evolving deep learning techniques have dominated critical fields such as natural language understanding,computer vision,multimodal processing and data mining,therefore greatly advancing the development of artificial intelligence(AI)technology.Among these advancements,transfer learning(TL)has emerged as a pivotal technique aimed at effectively reusing and sharing knowledge across multiple related models.This approach not only reduces the substantial costs associated with data collection and annotation,but also contributes to enhanced generalizability and capability of deep models.However,the exponential growth in the size,complexity,and depth of deep large models has presented serious challenges to traditional training and transfer algorithms,particularly in terms of computational and storage requirements.Such high computational complexity poses significant obstacles to effective knowledge transfer in resource-constrained scenarios,including but not limited to wearable technology,military applications,and healthcare systems.To address these challenges,efficient transfer learning algorithms have recently emerged as a promising solution,enabling agile adaptation and deployment of large models with minimal resource overhead.These algorithms are expected to become a key technological driver in the future development of AI.This paper stands out as the first comprehensive survey on the field of efficient transfer learning,aiming to systematically summarize research progress in this thriving research field over the past five years.Concretely,this paper investigates efficient transfer learning across three primary application fields:natural language processing,computer vision,and multimodal models.Among each application field,this paper further identifies and elaborates on five representative technical approaches that have gained promi-nence in recent research:modifying model structures,adjusting pre-training parameters,adapting original inputs(outputs),injecting adaptive parameters,and introducing adaptive modules.Each of these approaches is subjected to a comprehensive and thorough review,analyzing their respective strengths,limitations,and potential applications.This critical evaluation provides readers with a nuanced understanding of the current state of the art in efficient transfer learning.The primary contributions of this survey are threefold:(1)This survey presents the first systematic review of efficient transfer learning,offering invaluable technical insights and guidance for future research endeavors in this rapidly evolving field.(2)This survey proposes a novel technique-based frame-work that provides a clear and systematic research guideline,enabling readers to navigate the complex landscape of efficient transfer learning methodologies.(3)This survey conducts an in-depth analysis of the shortcomings and limitations of current methods,thereby identifying critical research gaps and providing insightful directions for future investigations.Efficient transfer learning serves as a crucial bridge between cutting-edge AI technologies and their practical applications in everyday life.It holds the potential to enable easier and cheaper access to the power of large models,benefiting a wide range of enterprises and individuals across various sectors.By providing comprehensive overviews of the current state of the art,solid theoretical foundations,practical guidance,along with critical insights into future research directions,this survey contributes significantly to the development of efficient transfer learning,and is hope to inspire researchers and practitioners to push the boundaries of the research field.

transfer learningdeep learningefficient methodmultimodal modellarge modellimited resources

李鑫尧、李晶晶、朱磊、申恒涛

展开 >

电子科技大学计算机科学与工程学院 成都 611731

同济大学电子信息与工程学院 上海 200092

迁移学习 深度学习 高效方法 多模态模型 大模型 资源受限

国家自然科学基金四川省自然科学基金TCL科技创新基金

621760422023NSFSC0483SS2024105

2024

计算机学报
中国计算机学会 中国科学院计算技术研究所

计算机学报

CSTPCD北大核心
影响因子:3.18
ISSN:0254-4164
年,卷(期):2024.47(11)