论文标题
重新思考深度学习中的不确定性:它是否以及如何改善鲁棒性
Rethinking Uncertainty in Deep Learning: Whether and How it Improves Robustness
论文作者
论文摘要
已知深层神经网络(DNN)容易受到对抗攻击,为此提出了许多补救措施。尽管对抗性训练(AT)被认为是最强大的防御,但在干净的例子和其他类型的攻击下,其表现不佳,例如攻击较大的扰动。同时,鼓励不确定产出的正规化器,例如熵最大化(ENTM)和标签平滑(LS)可以保持清洁示例的准确性,并在弱攻击下提高性能,但它们可以防御强烈攻击的能力仍然令人怀疑。在本文中,我们在对抗性学习领域中重新审视包括ENTM和LS在内的不确定性促进正规化器。我们表明,仅在小扰动下,单独的ENTM和LS才能提供鲁棒性。相反,我们表明,不确定性促进正规化器以原则上的方式进行补充,从而在干净的示例和各种攻击下都不断提高性能,尤其是具有较大扰动的攻击。我们进一步分析了不确定性促进正规化器如何从Jacobian矩阵$ \ nabla_x f(x;θ)$提高AT的性能,并发现ENTM有效地缩小了Jacobian矩阵的规范,从而促进了鲁棒性。
Deep neural networks (DNNs) are known to be prone to adversarial attacks, for which many remedies are proposed. While adversarial training (AT) is regarded as the most robust defense, it suffers from poor performance both on clean examples and under other types of attacks, e.g. attacks with larger perturbations. Meanwhile, regularizers that encourage uncertain outputs, such as entropy maximization (EntM) and label smoothing (LS) can maintain accuracy on clean examples and improve performance under weak attacks, yet their ability to defend against strong attacks is still in doubt. In this paper, we revisit uncertainty promotion regularizers, including EntM and LS, in the field of adversarial learning. We show that EntM and LS alone provide robustness only under small perturbations. Contrarily, we show that uncertainty promotion regularizers complement AT in a principled manner, consistently improving performance on both clean examples and under various attacks, especially attacks with large perturbations. We further analyze how uncertainty promotion regularizers enhance the performance of AT from the perspective of Jacobian matrices $\nabla_X f(X;θ)$, and find out that EntM effectively shrinks the norm of Jacobian matrices and hence promotes robustness.