论文标题
通过使用谓词题材结构增强训练句子来改善鲁棒性
Improving Robustness by Augmenting Training Sentences with Predicate-Argument Structures
论文作者
论文摘要
现有的NLP数据集包含各种偏见,并且模型往往会迅速学习这些偏见,从而又限制了它们的稳健性。现有的改善与数据集偏见的鲁棒性的方法主要集中在更改训练目标上,以便模型从偏见的示例中学到较少的东西。此外,它们主要集中在解决特定的偏见上,尽管它们改善了目标偏见的对抗评估集的性能,但它们可能会以其他方式偏向模型,因此会损害整体稳健性。在本文中,我们提议使用其相应的谓词题材结构来扩大培训数据中的输入句子,这些结构在不同的实现相同含义的不同实现方面提供了更高级别的抽象,并帮助模型识别句子的重要部分。我们表明,在没有针对特定偏见的情况下,我们的句子增强可以提高变压器模型对多种偏见的鲁棒性。此外,我们表明,即使训练数据不包含这种偏见,模型仍然容易受到词汇叠加偏差的影响,并且在这种情况下,句子的增加也可以提高鲁棒性。我们将在这种情况下发布我们的对抗数据集,以评估偏见,以及我们在https://github.com/ukplab/ukplab/data-aigmentation-for-robustness上的增强脚本。
Existing NLP datasets contain various biases, and models tend to quickly learn those biases, which in turn limits their robustness. Existing approaches to improve robustness against dataset biases mostly focus on changing the training objective so that models learn less from biased examples. Besides, they mostly focus on addressing a specific bias, and while they improve the performance on adversarial evaluation sets of the targeted bias, they may bias the model in other ways, and therefore, hurt the overall robustness. In this paper, we propose to augment the input sentences in the training data with their corresponding predicate-argument structures, which provide a higher-level abstraction over different realizations of the same meaning and help the model to recognize important parts of sentences. We show that without targeting a specific bias, our sentence augmentation improves the robustness of transformer models against multiple biases. In addition, we show that models can still be vulnerable to the lexical overlap bias, even when the training data does not contain this bias, and that the sentence augmentation also improves the robustness in this scenario. We will release our adversarial datasets to evaluate bias in such a scenario as well as our augmentation scripts at https://github.com/UKPLab/data-augmentation-for-robustness.