中国科学:信息科学(英文版)2024,Vol.67Issue(8) :202-216.DOI:10.1007/s11432-023-4016-0

Skill enhancement learning with knowledge distillation

Naijun LIU Fuchun SUN Bin FANG Huaping LIU
中国科学:信息科学(英文版)2024,Vol.67Issue(8) :202-216.DOI:10.1007/s11432-023-4016-0

Skill enhancement learning with knowledge distillation

Naijun LIU 1Fuchun SUN 1Bin FANG 1Huaping LIU1
扫码查看

作者信息

  • 1. Department of Computer Science and Technology,Tsinghua University,Beijing 100084,China
  • 折叠

Abstract

Skill learning through reinforcement learning has significantly progressed in recent years.How-ever,it often struggles to efficiently find optimal or near-optimal policies due to the inherent trial-and-error exploration in reinforcement learning.Although algorithms have been proposed to enhance skill learning efficacy,there is still much room for improvement in terms of skill learning performance and training sta-bility.In this paper,we propose an algorithm called skill enhancement learning with knowledge distillation(SELKD),which integrates multiple actors and multiple critics for skill learning.SELKD employs knowledge distillation to establish a mutual learning mechanism among actors.To mitigate critic overestimation bias,we introduce a novel target value calculation method.We also perform theoretical analysis to ensure the convergence of SELKD.Finally,experiments are conducted on several continuous control tasks,illustrating the effectiveness of the proposed algorithm.

Key words

skill learning/enhancement learning/reinforcement learning/knowledge distillation

引用本文复制引用

基金项目

"New Generation Artificial Intelligence"Key Field Research and Development Plan of Guangdong Province(2021B0101410002)

National Science and Technology Major Project of the Ministry of Science and Technology of China(2018AAA0102900)

National Natural Science Foundation of China(U22A2057)

National Natural Science Foundation of China(62133013)

出版年

2024
中国科学:信息科学(英文版)
中国科学院

中国科学:信息科学(英文版)

CSTPCDEI
影响因子:0.715
ISSN:1674-733X
段落导航相关论文