Robotics & Machine Learning Daily News2024,Issue(Feb.28) :44-45.DOI:10.1109/ACCESS.2024.3358408

New Robotics Research Reported from Kyonggi University (Applying Quantitative Model Checking to Analyze Safety in Reinforcement Learning)

Robotics & Machine Learning Daily News2024,Issue(Feb.28) :44-45.DOI:10.1109/ACCESS.2024.3358408

New Robotics Research Reported from Kyonggi University (Applying Quantitative Model Checking to Analyze Safety in Reinforcement Learning)

扫码查看

Abstract

Investigators publish new report on robotics. According to news reporting originating from Gyeonggi Do, South Korea, by NewsRx correspondents, research stated, "Reinforcement learning (RL) is rapidly used in safety-centric applications. However, many studies focus on generating optimal policy that achieves maximum rewards." Funders for this research include Institute of Information & Communications Technology Planning & Evaluation. Our news journalists obtained a quote from the research from Kyonggi University: "While maximum rewards are beneficial, safety constraints and non-functional requirements must also be considered in safety-centric applications to avoid dangerous situations. For example, in the case of food delivery robots in restaurants, RL should be used not only to find optimal policy that response to all customer requests through maximum rewards but also to consider safety constraints such as collision avoidance and nonfunctional requirements such as battery saving. In this paper, we investigated the fulfillment of safety constraints and non-functional requirements of learning models generated through RL with quantitative model checking. We experimented with various time steps and learning rates required for RL, targeting restaurant delivery robots. The functional requirement of these robots is to process all customer order requests, and the non-functional requirements are the number of steps and battery consumption to complete the task. Safety constraints include the amount of collision and the probability of collision. Through these experiments, we made three important findings. First, learning models that obtain maximum rewards may have a low degree of achievement of non-functional requirements and safety constraints. Second, as safety constraints are met, the degree of achievement of non-functional requirements may be low. Third, even if the maximum reward is not obtained, sacrificing non-functional requirements can maximize the achievement of safety constraints."

Key words

Kyonggi University/Gyeonggi Do/South Korea/Asia/Emerging Technologies/Machine Learning/Nano-robot/Reinforcement Learning/Robotics

引用本文复制引用

出版年

2024
Robotics & Machine Learning Daily News

Robotics & Machine Learning Daily News

ISSN:
参考文献量23
段落导航相关论文