论文标题
通过视觉不可感知的界限攻击不可抗拒的对抗防御
Attack Agnostic Adversarial Defense via Visual Imperceptible Bound
论文作者
论文摘要
深度学习算法对结构化和非结构化扰动的高敏感性促使了有效的对抗防御算法的发展。但是,现有防御算法缺乏普遍性以及不同数据库攻击算法的高度差异提出了有关国防算法有效性的几个问题。在这项研究中,我们旨在设计一种在一定范围内与可见和看不见的对抗性攻击之间具有牢固结合的防御模型。该界限与图像的视觉外观有关,我们将其称为\ textIt {Visual Unterction body(vib)}。为了计算此界限,我们提出了一种使用数据库特征的新方法。 VIB进一步用于衡量攻击算法的有效性。在包括C \&W($ L_2 $)和DeepFool在内的多次攻击中,对提出的防御模型的性能进行了评估。拟议的防御模型不仅能够提高针对多次攻击的鲁棒性,而且可以保留或提高原始清洁测试集中的分类精度。所提出的算法是攻击不可知论,即它不需要任何攻击算法的知识。
The high susceptibility of deep learning algorithms against structured and unstructured perturbations has motivated the development of efficient adversarial defense algorithms. However, the lack of generalizability of existing defense algorithms and the high variability in the performance of the attack algorithms for different databases raises several questions on the effectiveness of the defense algorithms. In this research, we aim to design a defense model that is robust within a certain bound against both seen and unseen adversarial attacks. This bound is related to the visual appearance of an image, and we termed it as \textit{Visual Imperceptible Bound (VIB)}. To compute this bound, we propose a novel method that uses the database characteristics. The VIB is further used to measure the effectiveness of attack algorithms. The performance of the proposed defense model is evaluated on the MNIST, CIFAR-10, and Tiny ImageNet databases on multiple attacks that include C\&W ($l_2$) and DeepFool. The proposed defense model is not only able to increase the robustness against several attacks but also retain or improve the classification accuracy on an original clean test set. The proposed algorithm is attack agnostic, i.e. it does not require any knowledge of the attack algorithm.