首页|New Robotics Research Reported from Kyonggi University (Applying Quantitative Model Checking to Analyze Safety in Reinforcement Learning)

New Robotics Research Reported from Kyonggi University (Applying Quantitative Model Checking to Analyze Safety in Reinforcement Learning)

扫码查看
Investigators publish new report on robotics. According to news reporting originating from Gyeonggi Do, South Korea, by NewsRx correspondents, research stated, "Reinforcement learning (RL) is rapidly used in safety-centric applications. However, many studies focus on generating optimal policy that achieves maximum rewards." Funders for this research include Institute of Information & Communications Technology Planning & Evaluation. Our news journalists obtained a quote from the research from Kyonggi University: "While maximum rewards are beneficial, safety constraints and non-functional requirements must also be considered in safety-centric applications to avoid dangerous situations. For example, in the case of food delivery robots in restaurants, RL should be used not only to find optimal policy that response to all customer requests through maximum rewards but also to consider safety constraints such as collision avoidance and nonfunctional requirements such as battery saving. In this paper, we investigated the fulfillment of safety constraints and non-functional requirements of learning models generated through RL with quantitative model checking. We experimented with various time steps and learning rates required for RL, targeting restaurant delivery robots. The functional requirement of these robots is to process all customer order requests, and the non-functional requirements are the number of steps and battery consumption to complete the task. Safety constraints include the amount of collision and the probability of collision. Through these experiments, we made three important findings. First, learning models that obtain maximum rewards may have a low degree of achievement of non-functional requirements and safety constraints. Second, as safety constraints are met, the degree of achievement of non-functional requirements may be low. Third, even if the maximum reward is not obtained, sacrificing non-functional requirements can maximize the achievement of safety constraints."

Kyonggi UniversityGyeonggi DoSouth KoreaAsiaEmerging TechnologiesMachine LearningNano-robotReinforcement LearningRobotics

2024

Robotics & Machine Learning Daily News

Robotics & Machine Learning Daily News

ISSN:
年,卷(期):2024.(Feb.28)
  • 23