论文标题
终身DP:终身机器学习中始终有界限的差异隐私
Lifelong DP: Consistently Bounded Differential Privacy in Lifelong Machine Learning
论文作者
论文摘要
在本文中,我们表明,不断学习新任务和记住先前任务的过程引入了未知的隐私风险和挑战以限制隐私损失。基于此,我们介绍了终身DP的正式定义,其中任何数据元组在越来越多的DP保护下都保护了任何数据元组在任何任务的训练集中受到保护,鉴于越来越多的任务流。始终如一的DP意味着只有一个固定值的DP隐私预算,而不管任务的数量多少。为了保留终身DP,我们提出了一种可扩展且异质的算法,称为L2DP-ML,具有流媒体批次培训,以有效地训练并继续释放L2M模型的新版本,鉴于数据尺寸的异质性和无需影响DP的DP保护设置的培训顺序,而无需影响私人培训设置。端到端的理论分析和彻底的评估表明,我们的机制明显好于保留终身DP的基线方法。 L2DP-ML的实现可在以下网址获得:https://github.com/haiphannjit/privatedeeplearning。
In this paper, we show that the process of continually learning new tasks and memorizing previous tasks introduces unknown privacy risks and challenges to bound the privacy loss. Based upon this, we introduce a formal definition of Lifelong DP, in which the participation of any data tuples in the training set of any tasks is protected, under a consistently bounded DP protection, given a growing stream of tasks. A consistently bounded DP means having only one fixed value of the DP privacy budget, regardless of the number of tasks. To preserve Lifelong DP, we propose a scalable and heterogeneous algorithm, called L2DP-ML with a streaming batch training, to efficiently train and continue releasing new versions of an L2M model, given the heterogeneity in terms of data sizes and the training order of tasks, without affecting DP protection of the private training set. An end-to-end theoretical analysis and thorough evaluations show that our mechanism is significantly better than baseline approaches in preserving Lifelong DP. The implementation of L2DP-ML is available at: https://github.com/haiphanNJIT/PrivateDeepLearning.