论文标题
混叠是对抗攻击的驱动力
Aliasing is a Driver of Adversarial Attacks
论文作者
论文摘要
混叠是信号处理中非常重要的概念,因为仔细考虑分辨率变化对于确保音频,图像和视频的传输质量的传输和处理质量至关重要。尽管如此,直到最近,Aliasing在深度学习中就很少考虑,所有共同的体系结构都不谨慎地进行亚采样,而无需考虑混叠效果。在这项工作中,我们调查了以下假设:对抗性扰动的存在部分是由于神经网络中的混叠。我们的最终目标是使用可解释的,未经训练的结构变化来提高对对抗性攻击的鲁棒性,这些变化仅源自混乱的第一原理。我们的贡献如下。首先,我们为一般图像转换建立了足够的条件。接下来,我们研究普通神经网络层中的混叠的来源,并从第一原则中得出简单的修改以消除或减少它。最后,我们的实验结果显示了抗叠缩和对抗攻击之间的扎实联系。简单地减少别名已经会导致更强大的分类器,并将反叠液与强大的训练结合在$ l_2 $攻击方面的强大训练,或者无用,或者对$ l _ {\ infty} $攻击的绩效造成的损失最小。
Aliasing is a highly important concept in signal processing, as careful consideration of resolution changes is essential in ensuring transmission and processing quality of audio, image, and video. Despite this, up until recently aliasing has received very little consideration in Deep Learning, with all common architectures carelessly sub-sampling without considering aliasing effects. In this work, we investigate the hypothesis that the existence of adversarial perturbations is due in part to aliasing in neural networks. Our ultimate goal is to increase robustness against adversarial attacks using explainable, non-trained, structural changes only, derived from aliasing first principles. Our contributions are the following. First, we establish a sufficient condition for no aliasing for general image transformations. Next, we study sources of aliasing in common neural network layers, and derive simple modifications from first principles to eliminate or reduce it. Lastly, our experimental results show a solid link between anti-aliasing and adversarial attacks. Simply reducing aliasing already results in more robust classifiers, and combining anti-aliasing with robust training out-performs solo robust training on $L_2$ attacks with none or minimal losses in performance on $L_{\infty}$ attacks.