重置(财务)
强化学习
计算机科学
任务(项目管理)
概率逻辑
课程
人工智能
自主代理人
控制器(灌溉)
风险分析(工程)
人机交互
机器学习
工程类
系统工程
生物
医学
金融经济学
教育学
经济
心理学
农学
作者
Matteo Turchetta,Andrey Kolobov,Shital Shah,Andreas Krause,Alekh Agarwal
出处
期刊:Cornell University - arXiv
日期:2020-01-01
被引量:38
标识
DOI:10.48550/arxiv.2006.12136
摘要
In safety-critical applications, autonomous agents may need to learn in an environment where mistakes can be very costly. In such settings, the agent needs to behave safely not only after but also while learning. To achieve this, existing safe reinforcement learning methods make an agent rely on priors that let it avoid dangerous situations during exploration with high probability, but both the probabilistic guarantees and the smoothness assumptions inherent in the priors are not viable in many scenarios of interest such as autonomous driving. This paper presents an alternative approach inspired by human teaching, where an agent learns under the supervision of an automatic instructor that saves the agent from violating constraints during learning. In this model, we introduce the monitor that neither needs to know how to do well at the task the agent is learning nor needs to know how the environment works. Instead, it has a library of reset controllers that it activates when the agent starts behaving dangerously, preventing it from doing damage. Crucially, the choices of which reset controller to apply in which situation affect the speed of agent learning. Based on observing agents' progress, the teacher itself learns a policy for choosing the reset controllers, a curriculum, to optimize the agent's final policy reward. Our experiments use this framework in two environments to induce curricula for safe and efficient learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI