论文标题
通过傅立叶口罩了解人工神经网络的鲁棒性和概括
Understanding robustness and generalization of artificial neural networks through Fourier masks
论文作者
论文摘要
尽管人工神经网络(ANN)在许多学科中取得了巨大的成功,但其计算的表征和关键属性(例如概括和鲁棒性)的起源仍然是开放的问题。最近的文献表明,具有良好概括属性的强大网络往往会偏向于处理图像中的低频。为了进一步探讨频率偏见假设,我们开发了一种算法,使我们能够学习调制面具,突出显示保留训练有素网络性能所需的基本输入频率。我们通过在输入频率中对此类调制的损失施加不变性来实现这一目标。我们首先使用我们的方法来测试经过对抗训练或数据增强网络的低频偏好假设。我们的结果表明,对抗性稳健的网络确实表现出低频偏见,但我们发现这种偏见也取决于频率空间的方向。但是,对于其他类型的数据增强不一定是正确的。我们的结果还表明,有效的基本频率最初是用于实现概括的基本频率。令人惊讶的是,通过这些调节面膜看到的图像不是可识别的,并且类似于纹理样的图案。
Despite the enormous success of artificial neural networks (ANNs) in many disciplines, the characterization of their computations and the origin of key properties such as generalization and robustness remain open questions. Recent literature suggests that robust networks with good generalization properties tend to be biased towards processing low frequencies in images. To explore the frequency bias hypothesis further, we develop an algorithm that allows us to learn modulatory masks highlighting the essential input frequencies needed for preserving a trained network's performance. We achieve this by imposing invariance in the loss with respect to such modulations in the input frequencies. We first use our method to test the low-frequency preference hypothesis of adversarially trained or data-augmented networks. Our results suggest that adversarially robust networks indeed exhibit a low-frequency bias but we find this bias is also dependent on directions in frequency space. However, this is not necessarily true for other types of data augmentation. Our results also indicate that the essential frequencies in question are effectively the ones used to achieve generalization in the first place. Surprisingly, images seen through these modulatory masks are not recognizable and resemble texture-like patterns.