论文标题
Lipschitz的边界和通过Laplacian平滑训练可证明训练
Lipschitz Bounds and Provably Robust Training by Laplacian Smoothing
论文作者
论文摘要
在这项工作中,我们提出了一个基于图形的学习框架,以训练具有对对抗性扰动的鲁棒性的模型。与基于正规化的方法相反,我们将对抗性鲁棒的学习问题提出,作为用Lipschitz限制的损失最小化之一,并表明相关Lagrangian的鞍点的特征是带有加权Laplace操作员的Poisson方程。此外,Lipschitz约束的Lagrange乘数给出了Laplace操作员的加权,这调节了最小化器对扰动的敏感性。然后,我们使用基于图的输入空间的基于图的离散化和一种原始的偶算法设计了一种可证明的强大训练方案,以收敛到Lagrangian的鞍点。我们的分析建立了具有约束性加权和对抗性学习的椭圆运算符之间的新联系。我们还研究了提高最小化器稳健性并损失的余量的互补问题,其损失幅度为Lipschitz常数的损失约束最小化问题。我们提出了一种技术来获得可靠的最小化器,并通过通过一系列梯度$ p $ norm最小化问题接近Lipschitz恒定最小化来评估基本Lipschitz下限。最终,我们的结果表明,对于所需的名义性能,存在着对对抗性扰动的敏感性的基本下限,这仅取决于损失函数和数据分布,并且超出该界限的鲁棒性的改善只能以名义性能为代价。事实证明,我们的培训计划在绩效和稳健性方面都达到了这些界限。
In this work we propose a graph-based learning framework to train models with provable robustness to adversarial perturbations. In contrast to regularization-based approaches, we formulate the adversarially robust learning problem as one of loss minimization with a Lipschitz constraint, and show that the saddle point of the associated Lagrangian is characterized by a Poisson equation with weighted Laplace operator. Further, the weighting for the Laplace operator is given by the Lagrange multiplier for the Lipschitz constraint, which modulates the sensitivity of the minimizer to perturbations. We then design a provably robust training scheme using graph-based discretization of the input space and a primal-dual algorithm to converge to the Lagrangian's saddle point. Our analysis establishes a novel connection between elliptic operators with constraint-enforced weighting and adversarial learning. We also study the complementary problem of improving the robustness of minimizers with a margin on their loss, formulated as a loss-constrained minimization problem of the Lipschitz constant. We propose a technique to obtain robustified minimizers, and evaluate fundamental Lipschitz lower bounds by approaching Lipschitz constant minimization via a sequence of gradient $p$-norm minimization problems. Ultimately, our results show that, for a desired nominal performance, there exists a fundamental lower bound on the sensitivity to adversarial perturbations that depends only on the loss function and the data distribution, and that improvements in robustness beyond this bound can only be made at the expense of nominal performance. Our training schemes provably achieve these bounds both under constraints on performance and~robustness.