论文标题

频lowcut池 - 插头与灾难性的过度拟合

FrequencyLowCut Pooling -- Plug & Play against Catastrophic Overfitting

论文作者

Grabinski, Julia, Jung, Steffen, Keuper, Janis, Keuper, Margret

论文摘要

在过去的几年中,卷积神经网络(CNN)一直是广泛的计算机视觉任务中的主导神经架构。从图像和信号处理的角度来看,大多数CNN的固有空间金字塔设计显然违反了基本的信号处理法,即在其下采样操作中采样定理。但是,由于抽样差似乎不影响模型的准确性,因此在模型鲁棒性开始受到更多关注之前,该问题已被广泛忽略。最近的工作[17]在对抗性攻击和分布变化的背景下,毕竟表明,CNN的脆弱性与不良的下降采样操作引起的混叠伪像之间存在很强的相关性。本文以这些发现为基础,并引入了一个混音的免费下采样操作,可以轻松地将其插入任何CNN体系结构:频lowcut池。我们的实验表明,与简单而快速的FGSM对抗训练结合使用,我们的超参数免费操作员显着提高了模型的鲁棒性,并避免了灾难性的过度拟合。

Over the last years, Convolutional Neural Networks (CNNs) have been the dominating neural architecture in a wide range of computer vision tasks. From an image and signal processing point of view, this success might be a bit surprising as the inherent spatial pyramid design of most CNNs is apparently violating basic signal processing laws, i.e. Sampling Theorem in their down-sampling operations. However, since poor sampling appeared not to affect model accuracy, this issue has been broadly neglected until model robustness started to receive more attention. Recent work [17] in the context of adversarial attacks and distribution shifts, showed after all, that there is a strong correlation between the vulnerability of CNNs and aliasing artifacts induced by poor down-sampling operations. This paper builds on these findings and introduces an aliasing free down-sampling operation which can easily be plugged into any CNN architecture: FrequencyLowCut pooling. Our experiments show, that in combination with simple and fast FGSM adversarial training, our hyper-parameter free operator significantly improves model robustness and avoids catastrophic overfitting.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源