论文标题
神经病原体:具有训练历史的深度神经网络的回归控制修复
NeuRecover: Regression-Controlled Repair of Deep Neural Networks with Training History
论文作者
论文摘要
鉴于对包括关键安全性的实用应用的需求不断增长,可以提高深神经网络(DNN)质量的系统技术至关重要。关键挑战来自更新DNN的几乎没有可控性。恢复以修复某些行为通常会对其他行为产生破坏性影响,从而导致回归,即更新的DNN失败,而原始行为正确处理了输入。当要求工程师调查针对安全或信任的密集保证活动中的失败时,此问题至关重要。基于搜索的DNN的维修技术有可能通过仅在DNN内的“负责任参数”上启用本地化更新来应对这一挑战。但是,尚未探索潜力来实现足够的可控性来抑制DNN维修任务中的回归。在本文中,我们提出了一种新型的DNN修复方法,该方法利用培训历史来判断哪种DNN参数应更改或不抑制回归。我们将该方法实施到称为Neurecover的工具中,并使用三个数据集对其进行了评估。我们的方法的表现超过了现有的方法,即通过少于四分之一,甚至在某些情况下是十分之一,即回归的数量。当维修要求紧密以解决特定的故障类型时,我们的方法特别有效。在这种情况下,我们的方法显示出较低的回归率(<2%),在许多情况下,这是重新培训引起的回归的十分之一。
Systematic techniques to improve quality of deep neural networks (DNNs) are critical given the increasing demand for practical applications including safety-critical ones. The key challenge comes from the little controllability in updating DNNs. Retraining to fix some behavior often has a destructive impact on other behavior, causing regressions, i.e., the updated DNN fails with inputs correctly handled by the original one. This problem is crucial when engineers are required to investigate failures in intensive assurance activities for safety or trust. Search-based repair techniques for DNNs have potentials to tackle this challenge by enabling localized updates only on "responsible parameters" inside the DNN. However, the potentials have not been explored to realize sufficient controllability to suppress regressions in DNN repair tasks. In this paper, we propose a novel DNN repair method that makes use of the training history for judging which DNN parameters should be changed or not to suppress regressions. We implemented the method into a tool called NeuRecover and evaluated it with three datasets. Our method outperformed the existing method by achieving often less than a quarter, even a tenth in some cases, number of regressions. Our method is especially effective when the repair requirements are tight to fix specific failure types. In such cases, our method showed stably low rates (<2%) of regressions, which were in many cases a tenth of regressions caused by retraining.