论文标题

对清晰行为的政策正规化

Policy Regularization for Legible Behavior

论文作者

Persiani, Michele, Hellström, Thomas

论文摘要

在加强学习中,可解释性通常意味着要深入了解代理机制,以便专家在检查时可以理解其决策。然而,通过文献的最终方法,对于在线环境中,交互的流畅性禁止对决策算法进行深入检查。为了支持在线设置中的可解释性,从可解释的计划文献方法中借用,通过在观察者模型中易于识别其意图,从而将重点放在代理商的可读性上。正如我们在本文中提出的那样,在代理策略中注入可清晰的行为不需要修改其学习算法的组件。相反,可以通过评估该政策如何产生观察者推断不正确的策略的观察结果来实现代理人的最佳政策以使其可读性。在我们的表述中,可透明度引入的决策边界会影响代理商政策返回在其他政策中具有很可能的行动的国家。在这些情况下,进行了这种行动与清晰/次优的行动之间的权衡。

In Reinforcement Learning interpretability generally means to provide insight into the agent's mechanisms such that its decisions are understandable by an expert upon inspection. This definition, with the resulting methods from the literature, may however fall short for online settings where the fluency of interactions prohibits deep inspections of the decision-making algorithm. To support interpretability in online settings it is useful to borrow from the Explainable Planning literature methods that focus on the legibility of the agent, by making its intention easily discernable in an observer model. As we propose in this paper, injecting legible behavior inside an agent's policy doesn't require modify components of its learning algorithm. Rather, the agent's optimal policy can be regularized for legibility by evaluating how the policy may produce observations that would make an observer infer an incorrect policy. In our formulation, the decision boundary introduced by legibility impacts the states in which the agent's policy returns an action that has high likelihood also in other policies. In these cases, a trade-off between such action, and legible/sub-optimal action is made.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源