论文标题

古哈特定律有例外吗?关于公平感知机器学习的道德理由

Are There Exceptions to Goodhart's Law? On the Moral Justification of Fairness-Aware Machine Learning

论文作者

Weerts, Hilde, Royakkers, Lambèr, Pechenizkiy, Mykola

论文摘要

公平感知的机器学习(FAIR-ML)技术是算法干预措施,旨在确保受机器学习模型预测影响的个人得到公平对待。这个问题通常被视为一个优化问题,其目的是在定量公平限制下实现高预测性能。但是,任何设计公平ML算法的尝试都必须假设一个世界法律有例外的世界:当公平度量成为优化限制时,它就不会停止成为一个好措施。在本文中,我们认为公平措施对古德哈特定律特别敏感。我们的主要贡献如下。首先,我们提出了关于公平指标理由的道德推理的框架。与现有工作相反,我们的框架结合了这样一种信念,即结果的分布是否公平,不仅取决于不平等的原因,还取决于决策主体必须获得特定利益或避免负担的道德主张。我们使用该框架来阐明道德和经验假设,在这些假设上,特定的公平度量指标与结果的公平分布相对应。其次,我们探讨采用公平指标作为公平算法中的约束的程度在道德上是合理的,这是由Hardt等人引入的Fair-ML算法所示例的。 (2016)。我们说明,通过公平的ML算法实施公平度量标准通常不会导致结果的公平分布,这些结果促使其使用,甚至可能损害个人旨在保护的个人。

Fairness-aware machine learning (fair-ml) techniques are algorithmic interventions designed to ensure that individuals who are affected by the predictions of a machine learning model are treated fairly. The problem is often posed as an optimization problem, where the objective is to achieve high predictive performance under a quantitative fairness constraint. However, any attempt to design a fair-ml algorithm must assume a world where Goodhart's law has an exception: when a fairness measure becomes an optimization constraint, it does not cease to be a good measure. In this paper, we argue that fairness measures are particularly sensitive to Goodhart's law. Our main contributions are as follows. First, we present a framework for moral reasoning about the justification of fairness metrics. In contrast to existing work, our framework incorporates the belief that whether a distribution of outcomes is fair, depends not only on the cause of inequalities but also on what moral claims decision subjects have to receive a particular benefit or avoid a burden. We use the framework to distil moral and empirical assumptions under which particular fairness metrics correspond to a fair distribution of outcomes. Second, we explore the extent to which employing fairness metrics as a constraint in a fair-ml algorithm is morally justifiable, exemplified by the fair-ml algorithm introduced by Hardt et al. (2016). We illustrate that enforcing a fairness metric through a fair-ml algorithm often does not result in the fair distribution of outcomes that motivated its use and can even harm the individuals the intervention was intended to protect.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源