论文标题
增强诱导的一致性正则分类
Augmentation-induced Consistency Regularization for Classification
论文作者
论文摘要
深度神经网络在许多监督的学习任务中都变得很流行,但是当培训数据集受到限制时,它们可能会因过度拟合而受苦。为了减轻这种情况,许多研究人员使用数据增强,这是一种广泛使用且有效的方法来增加数据集的种类。但是,数据增强引入的随机性会导致训练和推理之间不可避免的不一致,从而导致不良的进步。在本文中,我们提出了一个基于数据增强的一致性正则化框架,称为CR-AUG,该框架迫使数据增强生成的不同子模型的输出分布彼此一致。具体而言,CR-AUG评估了每个样品的两个增强版本的输出分布之间的差异,并且它利用了一个定型梯度操作来最大程度地减少一致性损失。我们实施CR-AUG来形象和音频分类任务,并进行广泛的实验,以验证其在提高分类器的概括能力方面的有效性。我们的CR-AUG框架是可用的,可以很容易地适应许多最新的网络体系结构。我们的经验结果表明,CR-AUG的表现优于基线方法,其余量很大。
Deep neural networks have become popular in many supervised learning tasks, but they may suffer from overfitting when the training dataset is limited. To mitigate this, many researchers use data augmentation, which is a widely used and effective method for increasing the variety of datasets. However, the randomness introduced by data augmentation causes inevitable inconsistency between training and inference, which leads to poor improvement. In this paper, we propose a consistency regularization framework based on data augmentation, called CR-Aug, which forces the output distributions of different sub models generated by data augmentation to be consistent with each other. Specifically, CR-Aug evaluates the discrepancy between the output distributions of two augmented versions of each sample, and it utilizes a stop-gradient operation to minimize the consistency loss. We implement CR-Aug to image and audio classification tasks and conduct extensive experiments to verify its effectiveness in improving the generalization ability of classifiers. Our CR-Aug framework is ready-to-use, it can be easily adapted to many state-of-the-art network architectures. Our empirical results show that CR-Aug outperforms baseline methods by a significant margin.