强化学习
计算机科学
概率逻辑
人工智能
模仿
钥匙(锁)
机器学习
钢筋
心理学
社会心理学
计算机安全
作者
Paweł Ładosz,Lilian Weng,Minwoo Kim,Hyondong Oh
标识
DOI:10.1016/j.inffus.2022.03.003
摘要
This paper reviews exploration techniques in deep reinforcement learning. Exploration techniques are of primary importance when solving sparse reward problems. In sparse reward problems, the reward is rare, which means that the agent will not find the reward often by acting randomly. In such a scenario, it is challenging for reinforcement learning to learn rewards and actions association. Thus more sophisticated exploration methods need to be devised. This review provides a comprehensive overview of existing exploration approaches, which are categorized based on the key contributions as follows reward novel states, reward diverse behaviours, goal-based methods, probabilistic methods, imitation-based methods, safe exploration and random-based methods. Then, the unsolved challenges are discussed to provide valuable future research directions. Finally, the approaches of different categories are compared in terms of complexity, computational effort and overall performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI