论文标题
认证黑框功能的跨域概括
Certifying Out-of-Domain Generalization for Blackbox Functions
论文作者
论文摘要
在有限的数据分布漂移下证明模型性能的鲁棒性最近引起了分布鲁棒性的范围的强烈兴趣。但是,现有技术要么对可以认证的模型类别和损失功能做出了强有力的假设,例如通过Lipschitz的梯度连续性表达的平滑度,要么需要解决复杂的优化问题。结果,这些技术的更广泛应用当前受其可扩展性和灵活性的限制 - 这些技术通常不会扩展到具有现代深神经网络的大规模数据集,或者无法处理可能不太平滑的损失功能,例如0-1损失。在本文中,我们着重于证明黑框模型和有限损失功能的分配鲁棒性的问题,并提出了基于Hellinger距离的新型认证框架。我们的认证技术缩放到Imagenet规模的数据集,复杂模型和各种损失功能。然后,我们专注于通过这种可伸缩性和灵活性来启用的一个特定应用程序,即,对大型神经网络和诸如准确性和AUC等大型神经网络的隔域概括进行认证。我们在许多数据集上实验验证了我们的认证方法,从ImageNet(从Imagenet)提供了第一个非易变认证的隔膜外概括到较小的分类任务,到我们能够与最先进的ART进行比较并表明我们的方法的性能更好。
Certifying the robustness of model performance under bounded data distribution drifts has recently attracted intensive interest under the umbrella of distributional robustness. However, existing techniques either make strong assumptions on the model class and loss functions that can be certified, such as smoothness expressed via Lipschitz continuity of gradients, or require to solve complex optimization problems. As a result, the wider application of these techniques is currently limited by its scalability and flexibility -- these techniques often do not scale to large-scale datasets with modern deep neural networks or cannot handle loss functions which may be non-smooth such as the 0-1 loss. In this paper, we focus on the problem of certifying distributional robustness for blackbox models and bounded loss functions, and propose a novel certification framework based on the Hellinger distance. Our certification technique scales to ImageNet-scale datasets, complex models, and a diverse set of loss functions. We then focus on one specific application enabled by such scalability and flexibility, i.e., certifying out-of-domain generalization for large neural networks and loss functions such as accuracy and AUC. We experimentally validate our certification method on a number of datasets, ranging from ImageNet, where we provide the first non-vacuous certified out-of-domain generalization, to smaller classification tasks where we are able to compare with the state-of-the-art and show that our method performs considerably better.