论文标题

Expanse:深度转移学习的深度 /渐进学习系统

EXPANSE: A Deep Continual / Progressive Learning System for Deep Transfer Learning

论文作者

Iman, Mohammadreza, Miller, John A., Rasheed, Khaled, Branch, Robert M., Arabnia, Hamid R.

论文摘要

深度转移学习技术试图通过重复使用知识来应对深度学习的局限性,对广泛的培训数据和培训成本的依赖。但是,当前的DTL技术遭受了灾难性的遗忘困难(失去了先前获得的知识)或在填充预训练模型或冻结预培训模型的一部分中的过度偏见的预培训模型(难以适应目标数据)。在DTL的子类别中,渐进式学习可以通过在冷冻预训练的模型的末尾添加新层来冻结早期层的效果。即使在许多情况下它已经成功,它仍无法处理遥远的源和目标数据。我们提出了一种新的持续/渐进学习方法,以深入转移学习来应对这些局限性。为了避免灾难性的遗忘和过度偏见的模型问题,我们通过在模型中扩展预训练的图层(向每一层添加新节点),而不仅仅是添加新图层,从而扩展了预训练的模型。因此,该方法被命名为Expanse。我们的实验结果证实,我们可以使用此技术来解决遥远的源和目标数据。同时,最终模型在源数据上仍然有效,实现了一种有希望的持续学习方法。此外,我们提供了一种培训受人体教育系统启发的深度学习模型的新方法。我们首先称赞了这一两步培训:学习基础知识,然后添加复杂性和不确定性。评估表明,两步训练提取了更有意义的特征和在误差表面上更精细的盆地,因为与常规训练相比,它可以达到更好的精度。 Expanse(模型扩展和两步培训)是一种适用于不同问题和DL模型的系统连续学习方法。

Deep transfer learning techniques try to tackle the limitations of deep learning, the dependency on extensive training data and the training costs, by reusing obtained knowledge. However, the current DTL techniques suffer from either catastrophic forgetting dilemma (losing the previously obtained knowledge) or overly biased pre-trained models (harder to adapt to target data) in finetuning pre-trained models or freezing a part of the pre-trained model, respectively. Progressive learning, a sub-category of DTL, reduces the effect of the overly biased model in the case of freezing earlier layers by adding a new layer to the end of a frozen pre-trained model. Even though it has been successful in many cases, it cannot yet handle distant source and target data. We propose a new continual/progressive learning approach for deep transfer learning to tackle these limitations. To avoid both catastrophic forgetting and overly biased-model problems, we expand the pre-trained model by expanding pre-trained layers (adding new nodes to each layer) in the model instead of only adding new layers. Hence the method is named EXPANSE. Our experimental results confirm that we can tackle distant source and target data using this technique. At the same time, the final model is still valid on the source data, achieving a promising deep continual learning approach. Moreover, we offer a new way of training deep learning models inspired by the human education system. We termed this two-step training: learning basics first, then adding complexities and uncertainties. The evaluation implies that the two-step training extracts more meaningful features and a finer basin on the error surface since it can achieve better accuracy in comparison to regular training. EXPANSE (model expansion and two-step training) is a systematic continual learning approach applicable to different problems and DL models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源