论文标题
元输入:如何利用现成的深神经网络
Meta Input: How to Leverage Off-the-Shelf Deep Neural Networks
论文作者
论文摘要
如今,尽管深层神经网络(DNN)在广泛的研究领域取得了显着的进步,但由于环境差异问题,它缺乏在现实世界应用中使用的适应性。这样的问题源于训练和测试环境之间的差异,并且众所周知,当验证的DNN模型应用于新的测试环境时,它会导致严重的性能退化。因此,在本文中,我们介绍了一种新颖的方法,该方法允许最终用户在自己的测试环境中利用预验证的DNN模型而无需修改模型。为此,我们提供了一个\ textIt {meta Input},这是一个进一步的输入,它转换了与培训数据相一致的测试数据的分布。仅通过考虑测试输入数据及其输出预测之间的关系,才可以通过少量测试数据来优化所提出的元输入。同样,它不需要任何了解网络的内部体系结构和重量参数的修改。然后,将获得的元输入添加到测试数据中,以将测试数据的分布转移到最初使用的培训数据的数据。结果,最终用户可以在自己的测试环境中利用训练有素的模型,这与培训环境不同。我们通过通过各种任务的全面实验来显示对环境差异的鲁棒性,从而验证了提出的元输入的有效性和多功能性。
These days, although deep neural networks (DNNs) have achieved a noticeable progress in a wide range of research area, it lacks the adaptability to be employed in the real-world applications because of the environment discrepancy problem. Such a problem originates from the difference between training and testing environments, and it is widely known that it causes serious performance degradation, when a pretrained DNN model is applied to a new testing environment. Therefore, in this paper, we introduce a novel approach that allows end-users to exploit pretrained DNN models in their own testing environment without modifying the models. To this end, we present a \textit{meta input} which is an additional input transforming the distribution of testing data to be aligned with that of training data. The proposed meta input can be optimized with a small number of testing data only by considering the relation between testing input data and its output prediction. Also, it does not require any knowledge of the network's internal architecture and modification of its weight parameters. Then, the obtained meta input is added to testing data in order to shift the distribution of testing data to that of originally used training data. As a result, end-users can exploit well-trained models in their own testing environment which can differ from the training environment. We validate the effectiveness and versatility of the proposed meta input by showing the robustness against the environment discrepancy through the comprehensive experiments with various tasks.