论文标题
双峰分布式二进制神经网络
Bimodal Distributed Binarized Neural Networks
论文作者
论文摘要
二元神经网络(BNN)是一种极为有希望的方法,可以大大降低深度神经网络的复杂性和功耗。然而,与其完整精确的对应物相比,二进化技术遭受了不合格的性能退化。 先前的工作主要集中在向前和向后相期间的符号函数近似策略上,以减少二进制过程中的量化误差。在这项工作中,我们提出了一种双模式分布式二进制方法(\ MethodName {})。这通过峰度正则化施加了网络权重的双模式分布。所提出的方法由我们称为模仿体重分布(WDM)的培训方案,该方案有效地模仿了其二进制二进制文件的全精度网络重量分布。在双纳拉化训练期间保留此分布会创建强大而有益的二元特征图,并大大减少BNN的概括误差。对CIFAR-10和ImageNet的广泛评估证明了我们方法比当前最新方案的优越性。我们的源代码,实验设置,培训日志和二进制模型可在\ url {https://github.com/blueanon/bd-bnn}上获得。
Binary Neural Networks (BNNs) are an extremely promising method to reduce deep neural networks' complexity and power consumption massively. Binarization techniques, however, suffer from ineligible performance degradation compared to their full-precision counterparts. Prior work mainly focused on strategies for sign function approximation during forward and backward phases to reduce the quantization error during the binarization process. In this work, we propose a Bi-Modal Distributed binarization method (\methodname{}). That imposes bi-modal distribution of the network weights by kurtosis regularization. The proposed method consists of a training scheme that we call Weight Distribution Mimicking (WDM), which efficiently imitates the full-precision network weight distribution to their binary counterpart. Preserving this distribution during binarization-aware training creates robust and informative binary feature maps and significantly reduces the generalization error of the BNN. Extensive evaluations on CIFAR-10 and ImageNet demonstrate the superiority of our method over current state-of-the-art schemes. Our source code, experimental settings, training logs, and binary models are available at \url{https://github.com/BlueAnon/BD-BNN}.