论文标题
对深度学习反法敏化的对抗性示例的增加
Increased-confidence adversarial examples for deep learning counter-forensics
论文作者
论文摘要
对抗性示例的可转移性是在现实生活中基于深度学习(DL)的多媒体取证(MMF)技术的这种攻击的关键问题。实际上,在攻击者没有完全了解受攻击的系统的情况下,对抗性示例的可转移性实际上将为部署成功的反法医攻击开辟道路。一些初步作品表明,至少在采用了最受欢迎的库中实施的攻击的基本版本时,针对基于CNN的图像取证探测器的对抗性示例通常是不可转让的。在本文中,我们介绍了一种一般策略,以提高攻击的强度并评估其在这种强度变化时的转移性。我们从实验上表明,以这种方式,可以大大提高攻击性,而牺牲更大的失真为代价。我们的研究证实了即使在多媒体取证方案中也存在对抗性示例所带来的安全威胁,因此呼吁采取新的防御策略来提高基于DL的MMF技术的安全性。
Transferability of adversarial examples is a key issue to apply this kind of attacks against multimedia forensics (MMF) techniques based on Deep Learning (DL) in a real-life setting. Adversarial example transferability, in fact, would open the way to the deployment of successful counter forensics attacks also in cases where the attacker does not have a full knowledge of the to-be-attacked system. Some preliminary works have shown that adversarial examples against CNN-based image forensics detectors are in general non-transferrable, at least when the basic versions of the attacks implemented in the most popular libraries are adopted. In this paper, we introduce a general strategy to increase the strength of the attacks and evaluate their transferability when such a strength varies. We experimentally show that, in this way, attack transferability can be largely increased, at the expense of a larger distortion. Our research confirms the security threats posed by the existence of adversarial examples even in multimedia forensics scenarios, thus calling for new defense strategies to improve the security of DL-based MMF techniques.