论文标题
分心是您公平所需要的
Distraction is All You Need for Fairness
论文作者
论文摘要
培训数据集中的偏见必须针对分类任务的各个组进行管理,以确保均等或平等待遇。随着人工智能模型的最新增长及其在自动决策中的不断扩展,确保这些模型没有偏见至关重要。有大量证据表明,这些模型可以包含甚至扩大对其目标功能和学习算法固有的数据中存在的偏差;许多研究人员将注意力转移到不同的方向上,即,将数据变化为统计独立,对抗性培训,以限制旨在最大化平价等特定竞争者的能力。这些方法会导致信息丢失,并不能在准确性和公平性之间提供适当的平衡,或者不能确保限制培训中的偏见。为此,我们提出了一种强大的策略,用于训练称为“干扰模块”的深度学习模型,理论上可以证明可以有效控制影响分类结果的偏见。该方法可以使用不同的数据类型(例如表格,图像,图形等)使用。我们通过对UCI成人和Heritage Health数据集(表格),POKEC-Z,POKEC-N和NBA数据集(Graph)以及Celeba DataSet(Vision)进行测试来证明该方法的效力。使用每个数据集的公平文献中提出的最新方法,我们展示的模型优于这些提出的方法,可以最大程度地减少偏见和保持准确性。
Bias in training datasets must be managed for various groups in classification tasks to ensure parity or equal treatment. With the recent growth in artificial intelligence models and their expanding role in automated decision-making, ensuring that these models are not biased is vital. There is an abundance of evidence suggesting that these models could contain or even amplify the bias present in the data on which they are trained, inherent to their objective function and learning algorithms; Many researchers direct their attention to this issue in different directions, namely, changing data to be statistically independent, adversarial training for restricting the capabilities of a particular competitor who aims to maximize parity, etc. These methods result in information loss and do not provide a suitable balance between accuracy and fairness or do not ensure limiting the biases in training. To this end, we propose a powerful strategy for training deep learning models called the Distraction module, which can be theoretically proven effective in controlling bias from affecting the classification results. This method can be utilized with different data types (e.g., Tabular, images, graphs, etc.). We demonstrate the potency of the proposed method by testing it on UCI Adult and Heritage Health datasets (tabular), POKEC-Z, POKEC-N and NBA datasets (graph), and CelebA dataset (vision). Using state-of-the-art methods proposed in the fairness literature for each dataset, we exhibit our model is superior to these proposed methods in minimizing bias and maintaining accuracy.