论文标题
深度强化学习中的探索:一项调查
Exploration in Deep Reinforcement Learning: A Survey
论文作者
论文摘要
本文回顾了深度强化学习中的探索技术。在解决稀疏奖励问题时,探索技术至关重要。在稀疏的奖励问题中,奖励很少见,这意味着代理商不会通过随机行动来经常找到奖励。在这种情况下,强化学习学习奖励和行动协会是一项挑战。因此,需要设计更复杂的探索方法。这篇综述提供了现有探索方法的全面概述,这些方法是根据关键贡献进行分类的,如下所示,奖励新颖的状态,奖励各种行为,基于目标的方法,概率方法,基于模仿的方法,安全探索和基于随机的方法。然后,讨论了未解决的挑战,以提供宝贵的未来研究指示。最后,根据复杂性,计算工作和整体绩效比较不同类别的方法。
This paper reviews exploration techniques in deep reinforcement learning. Exploration techniques are of primary importance when solving sparse reward problems. In sparse reward problems, the reward is rare, which means that the agent will not find the reward often by acting randomly. In such a scenario, it is challenging for reinforcement learning to learn rewards and actions association. Thus more sophisticated exploration methods need to be devised. This review provides a comprehensive overview of existing exploration approaches, which are categorized based on the key contributions as follows reward novel states, reward diverse behaviours, goal-based methods, probabilistic methods, imitation-based methods, safe exploration and random-based methods. Then, the unsolved challenges are discussed to provide valuable future research directions. Finally, the approaches of different categories are compared in terms of complexity, computational effort and overall performance.