论文标题
学习在部分已知的环境中跟踪动态目标
Learning to Track Dynamic Targets in Partially Known Environments
论文作者
论文摘要
我们使用深入的增强学习方法(RL)方法解决了主动目标跟踪,这是自主系统中的重要任务之一。在此问题中,自主代理的任务是使用其机载传感器获取有关兴趣目标的信息。这个问题中的经典挑战是系统模型依赖性和计算信息理论成本功能在漫长计划范围内的难度。 RL为这些挑战提供了解决方案,因为其有效计划范围的长度不会影响计算复杂性,并且它降低了算法对系统模型的强烈依赖性。特别是,我们引入了主动跟踪目标网络(ATTN),这是一种统一的RL策略,能够求解主动目标跟踪的主要子任务 - 视野跟踪,导航和探索。该政策显示出具有部分已知目标模型的敏捷和异常目标的鲁棒行为。此外,相同的策略能够在障碍环境中导航以达到遥远的目标,并在目标位于意外地点时探索环境。
We solve active target tracking, one of the essential tasks in autonomous systems, using a deep reinforcement learning (RL) approach. In this problem, an autonomous agent is tasked with acquiring information about targets of interests using its onboard sensors. The classical challenges in this problem are system model dependence and the difficulty of computing information-theoretic cost functions for a long planning horizon. RL provides solutions for these challenges as the length of its effective planning horizon does not affect the computational complexity, and it drops the strong dependency of an algorithm on system models. In particular, we introduce Active Tracking Target Network (ATTN), a unified RL policy that is capable of solving major sub-tasks of active target tracking -- in-sight tracking, navigation, and exploration. The policy shows robust behavior for tracking agile and anomalous targets with a partially known target model. Additionally, the same policy is able to navigate in obstacle environments to reach distant targets as well as explore the environment when targets are positioned in unexpected locations.