首页|通用大模型演进路线

通用大模型演进路线

扫码查看
随着人工智能技术的飞速发展,通用大模型(GLMs)已经成为人工智能领域的重要研究方向.通用大模型拥有超大规模参数,通过大规模数据进行训练,具备强大的学习和推理能力.这些模型在自然语言处理、图像识别、代码生成等多种任务中展现出卓越的能力.回顾了通用大模型的发展历程,梳理关键技术节点,从早期基于规则的系统和传统机器学习模型,到深度学习的崛起,再到Transformer架构,以及GPT系列及国内外通用大模型的进展.尽管GLMs在多个领域取得了显著进展,但其发展也面临诸多挑战,包括计算资源需求、数据偏见与伦理问题及模型的解释性与透明性.分析了这些挑战,并探讨了GLMs未来发展的5个关键方向:模型优化、多模态学习、具情感大模型、数据与知识双驱动以及伦理与社会影响.通过这些策略,通用大模型有望在未来实现更广泛和深入的应用,推动人工智能技术的持续进步.
Evolution of general large models
With the rapid development of artificial intelligence(AI)technology,general large models(GLMs)have become a significant research focus in the AI field.GLMs typically possess an extensive number of parameters,are trained on massive datasets and exhibit robust learning and reasoning capabilities.These models demonstrate outstanding performance in various tasks,including natural language processing,image recognition,and code generation.This paper reviews the evolution of GLMs and the key technology nodes,from the early rule-based systems and traditional machine learning models to the rise of deep learning,the introduction of the Transformer architecture,and the advancements in the GPT series and other GLMS over the world.Despite the significant progress,GLMs face numerous challenges,such as high computational resource demands,data bias,ethical issues,and model interpretability and transparency.This paper analyzes these challenges and explores five key future development directions for GLMs:model optimization,multimodal learning,emotionally intelligent models,data and knowledge dual-driven models,and ethical and societal impacts.By adopting these strategies,GLMs are expected to achieve broader and deeper applications,driving continuous progress in AI technology.

general large modelsartificial intelligencedeep learningtransformer architectureGPT series

任福继、张彦如

展开 >

电子科技大学计算机科学与工程学院,成都 611731

电子科技大学(深圳)高等研究院,深圳 518110

通用大模型 人工智能 深度学习 Transformer架构 GPT系列

2024

科技导报
中国科学技术协会

科技导报

CSTPCD北大核心
影响因子:0.559
ISSN:1000-7857
年,卷(期):2024.42(12)
  • 32