论文标题
基于自我训练的域适应性的两相伪标记致密性
Two-phase Pseudo Label Densification for Self-training based Domain Adaptation
论文作者
论文摘要
最近,深层的自我训练方法是对无监督领域适应的有力解决方案。自我训练方案涉及目标数据的迭代处理;它生成目标伪标签并重新培训网络。但是,由于仅将自信的预测视为伪标签,因此现有的自我训练方法不可避免地会在实践中产生稀疏的伪标签。我们认为这至关重要,因为由此产生的训练信号不足会导致次优,容易出错的模型。为了解决这个问题,我们提出了一种新型的两相伪标签致密框架,称为TPLD。在第一阶段,我们使用滑动窗口投票来传播自信的预测,并利用图像中的固有空间相关。在第二阶段,我们执行基于置信的易于硬性分类。对于简单的样品,我们现在使用他们的完整伪标签。对于艰难的人来说,我们采用对抗性学习来实施难以满意的功能一致性。为了简化训练过程并避免嘈杂的预测,我们将自举机制介绍给原始的自我训练损失。我们表明,所提出的TPLD可以轻松地集成到现有的基于自我训练的方法中,并显着改善性能。结合最近提出的CRST自我训练框架,我们在两个标准的UDA基准测试中获得了新的最新结果。
Recently, deep self-training approaches emerged as a powerful solution to the unsupervised domain adaptation. The self-training scheme involves iterative processing of target data; it generates target pseudo labels and retrains the network. However, since only the confident predictions are taken as pseudo labels, existing self-training approaches inevitably produce sparse pseudo labels in practice. We see this is critical because the resulting insufficient training-signals lead to a suboptimal, error-prone model. In order to tackle this problem, we propose a novel Two-phase Pseudo Label Densification framework, referred to as TPLD. In the first phase, we use sliding window voting to propagate the confident predictions, utilizing intrinsic spatial-correlations in the images. In the second phase, we perform a confidence-based easy-hard classification. For the easy samples, we now employ their full pseudo labels. For the hard ones, we instead adopt adversarial learning to enforce hard-to-easy feature alignment. To ease the training process and avoid noisy predictions, we introduce the bootstrapping mechanism to the original self-training loss. We show the proposed TPLD can be easily integrated into existing self-training based approaches and improves the performance significantly. Combined with the recently proposed CRST self-training framework, we achieve new state-of-the-art results on two standard UDA benchmarks.