论文标题

力量自适应对抗训练

Strength-Adaptive Adversarial Training

论文作者

Yu, Chaojian, Zhou, Dawei, Shen, Li, Yu, Jun, Han, Bo, Gong, Mingming, Wang, Nannan, Liu, Tongliang

论文摘要

事实证明,对抗性训练(AT)可靠地改善网络对对抗数据的鲁棒性。但是,当前使用预先指定的扰动预算在学习强大的网络方面存在局限性。首先,将预先指定的扰动预算应用于各种模型能力的网络上,将产生自然和鲁棒精度之间的鲁棒性差异,从而偏离了稳健网络的desideratum。其次,随着网络鲁棒性的增长,受预先指定的扰动预算约束的对抗性训练数据的攻击强度无法升级,这会导致强大的过度拟合并进一步降低对抗性鲁棒性。为了克服这些局限性,我们建议\ emph {强度自适应对抗训练}(SAAT)。具体而言,对手采用对抗性损失约束来生成对抗性训练数据。在此限制下,扰动预算将根据对抗数据的训练状态进行自适应调整,这可以有效地避免强大的过度拟合。此外,SAAT通过对抗性损失明确限制了训练数据的攻击强度,该损失可以操纵训练期间的模型能力调度,从而可以灵活地控制稳健性差异的程度,并调整自然准确性和鲁棒性之间的权衡。广泛的实验表明,我们的提议增强了对抗训练的鲁棒性。

Adversarial training (AT) is proved to reliably improve network's robustness against adversarial data. However, current AT with a pre-specified perturbation budget has limitations in learning a robust network. Firstly, applying a pre-specified perturbation budget on networks of various model capacities will yield divergent degree of robustness disparity between natural and robust accuracies, which deviates from robust network's desideratum. Secondly, the attack strength of adversarial training data constrained by the pre-specified perturbation budget fails to upgrade as the growth of network robustness, which leads to robust overfitting and further degrades the adversarial robustness. To overcome these limitations, we propose \emph{Strength-Adaptive Adversarial Training} (SAAT). Specifically, the adversary employs an adversarial loss constraint to generate adversarial training data. Under this constraint, the perturbation budget will be adaptively adjusted according to the training state of adversarial data, which can effectively avoid robust overfitting. Besides, SAAT explicitly constrains the attack strength of training data through the adversarial loss, which manipulates model capacity scheduling during training, and thereby can flexibly control the degree of robustness disparity and adjust the tradeoff between natural accuracy and robustness. Extensive experiments show that our proposal boosts the robustness of adversarial training.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源