论文标题
跨函数的联合线性和非线性计算,以有效地保护隐私神经网络推断
Joint Linear and Nonlinear Computation across Functions for Efficient Privacy-Preserving Neural Network Inference
论文作者
论文摘要
虽然令人鼓舞的是,目的是目睹隐私机器学习作为一项服务(MLAA)的最新发展,但仍存在其在现实世界应用中部署的显着绩效差距。我们观察到最先进的框架遵循每个功能输出的计算和共享原理,在线函数中的总和是两个函数输出的两个步骤中的最后一个,涉及所有旋转(这是他操作最昂贵),而非线性功能中的多路复用,这也是函数的两个步骤的函数输出的最后一个步骤,引入了可互动的通信。因此,我们挑战常规的计算和共享逻辑,并跨功能引入第一个关节线性和非线性计算,该功能以1)计算非线性函数的PHE三重态,从而消除了多路复用; 2)矩阵编码以计算线性函数,从而消除了所有用于求和的旋转; 3)网络改编以重新组装模型结构,并尽可能利用联合计算模块。通过数值复杂性来验证增强效率,实验证明了最新模型中使用的各种功能以及主流神经网络上最高可达5倍的速度。
While it is encouraging to witness the recent development in privacy-preserving Machine Learning as a Service (MLaaS), there still exists a significant performance gap for its deployment in real-world applications. We observe the state-of-the-art frameworks follow a compute-and-share principle for every function output where the summing in linear functions, which is the last of two steps for function output, involves all rotations (which is the most expensive HE operation), and the multiplexing in nonlinear functions, which is also the last of two steps for function output, introduces noticeable communication rounds. Therefore, we challenge the conventional compute-and-share logic and introduce the first joint linear and nonlinear computation across functions that features by 1) the PHE triplet for computing the nonlinear function, with which the multiplexing is eliminated; 2) the matrix encoding to calculate the linear function, with which all rotations for summing is removed; and 3) the network adaptation to reassemble the model structure, with which the joint computation module is utilized as much as possible. The boosted efficiency is verified by the numerical complexity, and the experiments demonstrate up to 13x speedup for various functions used in the state-of-the-art models and up to 5x speedup over mainstream neural networks.