论文标题
在不断发展的数据流中更改局部解释性检测
Change Detection for Local Explainability in Evolving Data Streams
论文作者
论文摘要
随着复杂的机器学习模型越来越多地用于银行,交易或信用评分等敏感应用程序,因此对可靠的解释机制的需求不断增长。局部特征归因方法已成为事后和模型不合时式解释的流行技术。但是,归因方法通常假设一个固定环境,在该环境中,对预测模型进行了训练并保持稳定。结果,通常不清楚本地归因在现实,不断发展的设置(例如流媒体和在线应用程序)中的行为。在本文中,我们讨论了时间变化对本地特征归因的影响。特别是,我们表明,每次更新预测模型或概念漂移都会改变数据生成分布时,本地归因都会变得过时。因此,数据流中的本地特征归因只有在将我们能够随着时间的推移随时间推移检测并响应局部变化的机制结合使用时才能提供较高的解释力。为此,我们提出了一个clededs,这是一个灵活的模型无关框架,用于检测局部变化和概念漂移。 CDEREDS是基于归因的解释技术的直观扩展,以识别过时的本地属性并实现了更具针对性的重新计算。在实验中,我们还表明,所提出的框架可以可靠地检测到本地和全球概念漂移。因此,我们的工作在在线机器学习中有助于更有意义,更强大的解释性。
As complex machine learning models are increasingly used in sensitive applications like banking, trading or credit scoring, there is a growing demand for reliable explanation mechanisms. Local feature attribution methods have become a popular technique for post-hoc and model-agnostic explanations. However, attribution methods typically assume a stationary environment in which the predictive model has been trained and remains stable. As a result, it is often unclear how local attributions behave in realistic, constantly evolving settings such as streaming and online applications. In this paper, we discuss the impact of temporal change on local feature attributions. In particular, we show that local attributions can become obsolete each time the predictive model is updated or concept drift alters the data generating distribution. Consequently, local feature attributions in data streams provide high explanatory power only when combined with a mechanism that allows us to detect and respond to local changes over time. To this end, we present CDLEEDS, a flexible and model-agnostic framework for detecting local change and concept drift. CDLEEDS serves as an intuitive extension of attribution-based explanation techniques to identify outdated local attributions and enable more targeted recalculations. In experiments, we also show that the proposed framework can reliably detect both local and global concept drift. Accordingly, our work contributes to a more meaningful and robust explainability in online machine learning.