论文标题
但是元学习可以快速适应,也可以轻松破裂
Yet Meta Learning Can Adapt Fast, It Can Also Break Easily
论文作者
论文摘要
元学习算法已被广泛应用于许多任务以进行有效的学习,例如少量图像分类和快速增强学习。在元培训期间,元学习者从各种学习任务中制定了一种共同的学习策略或经验。因此,在元测试期间,即使有一些培训样本,元学习者也可以使用学习的策略快速适应新任务。但是,从可靠性和鲁棒性方面,元学习仍然存在一个黑暗的一面。特别是,元学习容易受到对抗攻击的影响?换句话说,如果对手无情地操纵给定的训练集,那么训练有素的元学习者是否会利用其学到的经验来建立错误或可能无用的知识?在没有理解此问题的情况下,将元学习应用于安全至关重要的应用是极大的风险。因此,在本文中,我们进行了有关在少量拍摄分类问题下对元学习的对抗性攻击的初步研究。特别是,我们正式定义了元学习独有的对抗攻击的关键要素,并提出了在各种环境下针对元学习的第一种攻击算法。我们评估了提出的攻击策略的有效性以及几种代表性元学习算法的鲁棒性。实验结果表明,提出的攻击策略可以轻松打破元学习者,元学习易受对抗性攻击的影响。拟议框架的实施将在接受本文后发布。
Meta learning algorithms have been widely applied in many tasks for efficient learning, such as few-shot image classification and fast reinforcement learning. During meta training, the meta learner develops a common learning strategy, or experience, from a variety of learning tasks. Therefore, during meta test, the meta learner can use the learned strategy to quickly adapt to new tasks even with a few training samples. However, there is still a dark side about meta learning in terms of reliability and robustness. In particular, is meta learning vulnerable to adversarial attacks? In other words, would a well-trained meta learner utilize its learned experience to build wrong or likely useless knowledge, if an adversary unnoticeably manipulates the given training set? Without the understanding of this problem, it is extremely risky to apply meta learning in safety-critical applications. Thus, in this paper, we perform the initial study about adversarial attacks on meta learning under the few-shot classification problem. In particular, we formally define key elements of adversarial attacks unique to meta learning and propose the first attacking algorithm against meta learning under various settings. We evaluate the effectiveness of the proposed attacking strategy as well as the robustness of several representative meta learning algorithms. Experimental results demonstrate that the proposed attacking strategy can easily break the meta learner and meta learning is vulnerable to adversarial attacks. The implementation of the proposed framework will be released upon the acceptance of this paper.