论文标题
在标签噪声下进行有效学习的嘈杂并发培训
Noisy Concurrent Training for Efficient Learning under Label Noise
论文作者
论文摘要
深度神经网络(DNNS)无法在标签噪声下有效学习,并已被证明可以记住影响其概括性能的随机标签。我们考虑孤立地学习,使用一式编码标签作为唯一的监督来源,而缺乏正规化来阻止记忆是标准培训程序的主要缺点。因此,我们提出了嘈杂的并发培训(NCT),该培训利用协作学习将两个模型之间的共识用作附加的监督来源。此外,受到大脑试验性变异性的启发,我们提出了一种反直觉的正则化技术,目标变异性需要随机更改每批培训样本的百分比标签,以阻止DNNS中的记忆和过度生殖。目标变异性独立应用于每个模型,以使其差异并避免确认偏差。由于DNN倾向于在记住嘈杂标签之前首先优先考虑学习简单模式,因此我们采用了动态学习方案,随着培训的进行,这两个模型越来越多地依赖他们的共识。 NCT还逐步提高了目标变异性,以避免在后期阶段记忆。我们证明了我们的方法对合成和现实世界噪声基准数据集的有效性。
Deep neural networks (DNNs) fail to learn effectively under label noise and have been shown to memorize random labels which affect their generalization performance. We consider learning in isolation, using one-hot encoded labels as the sole source of supervision, and a lack of regularization to discourage memorization as the major shortcomings of the standard training procedure. Thus, we propose Noisy Concurrent Training (NCT) which leverages collaborative learning to use the consensus between two models as an additional source of supervision. Furthermore, inspired by trial-to-trial variability in the brain, we propose a counter-intuitive regularization technique, target variability, which entails randomly changing the labels of a percentage of training samples in each batch as a deterrent to memorization and over-generalization in DNNs. Target variability is applied independently to each model to keep them diverged and avoid the confirmation bias. As DNNs tend to prioritize learning simple patterns first before memorizing the noisy labels, we employ a dynamic learning scheme whereby as the training progresses, the two models increasingly rely more on their consensus. NCT also progressively increases the target variability to avoid memorization in later stages. We demonstrate the effectiveness of our approach on both synthetic and real-world noisy benchmark datasets.