论文标题

沟通效率的联合学习的三元压缩

Ternary Compression for Communication-Efficient Federated Learning

论文作者

Xu, Jinjin, Du, Wenli, Cheng, Ran, He, Wangli, Jin, Yaochu

论文摘要

在许多现实世界应用中,对存储在不同位置的大量数据进行学习是必不可少的。但是,由于智能移动设备和IoT设备的使用日益增长的使用,共享数据的挑战充满了挑战。联合学习通过共同培训全球模型,而无需将分布在多个设备上的数据上载到中央服务器上,从而为隐私保护和安全的机器学习提供了潜在的解决方案。但是,大多数现有的联邦学习工作都采用具有完整精确权重的机器学习模型,几乎所有这些模型都包含大量冗余参数,这些参数无需传输到服务器,从而消耗了过多的通信成本。为了解决这个问题,我们提出了一种联合训练的三元量化(FTTQ)算法,该算法通过自学习量化因子优化客户端的量化网络。给出了量化因子收敛性,FTTQ无偏的理论证明以及减轻的权重差异。在FTTQ的基础上,我们提出了一个三元联合平均协议(T-FEDAVG),以减少联合学习系统的上游和下游通信。进行了经验实验,以培训广泛使用的深度学习模型,以公开可用的数据集进行培训,我们的结果表明,与规范的联邦学习算法相比,提议的T-FEDAVG有效地降低了沟通成本,甚至可以在非IID数据上实现稍微更好的性能。

Learning over massive data stored in different locations is essential in many real-world applications. However, sharing data is full of challenges due to the increasing demands of privacy and security with the growing use of smart mobile devices and IoT devices. Federated learning provides a potential solution to privacy-preserving and secure machine learning, by means of jointly training a global model without uploading data distributed on multiple devices to a central server. However, most existing work on federated learning adopts machine learning models with full-precision weights, and almost all these models contain a large number of redundant parameters that do not need to be transmitted to the server, consuming an excessive amount of communication costs. To address this issue, we propose a federated trained ternary quantization (FTTQ) algorithm, which optimizes the quantized networks on the clients through a self-learning quantization factor. Theoretical proofs of the convergence of quantization factors, unbiasedness of FTTQ, as well as a reduced weight divergence are given. On the basis of FTTQ, we propose a ternary federated averaging protocol (T-FedAvg) to reduce the upstream and downstream communication of federated learning systems. Empirical experiments are conducted to train widely used deep learning models on publicly available datasets, and our results demonstrate that the proposed T-FedAvg is effective in reducing communication costs and can even achieve slightly better performance on non-IID data in contrast to the canonical federated learning algorithms.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源