论文标题
重新思考从ANN到SNNS的桥梁进行训练
Rethinking Pretraining as a Bridge from ANNs to SNNs
论文作者
论文摘要
尖峰神经网络(SNN)被称为一种典型的脑启发模型,其独特的神经元动力学,各种编码方案和低功耗特性的独特特征。如何获得高准确模型一直是SNN领域的主要挑战。当前,有两种主流方法,即通过将训练有素的人工神经网络(ANN)转换为SNN对应物或直接训练SNN来获得转换后的SNN。但是,转换后的SNN的推理时间太长,而SNN训练通常非常昂贵且效率低下。在这项工作中,提出了一种新的SNN培训范式,通过将两种不同的训练方法的概念与前训练技术和基于BP的深SNN训练机制相结合。我们认为,拟议的范式是培训SNN的更有效的管道。管道包括用于静态数据传输任务的管道和动态数据传输任务的管道。 SOTA结果是在大规模事件驱动的数据集ES-IMAGENET中获得的。对于训练加速,我们使用Imagenet-1K上的1/10培训时间和在ES-Imagenet上的1/10培训时间获得了与类似的LIF-SNN相同(或更高)的最佳精度,还为新的数据集ES-UCF101提供了时间准确的基准。这些实验结果揭示了ANN和SNN之间参数功能的相似性,还证明了该SNN训练管道的各种潜在应用。
Spiking neural networks (SNNs) are known as a typical kind of brain-inspired models with their unique features of rich neuronal dynamics, diverse coding schemes and low power consumption properties. How to obtain a high-accuracy model has always been the main challenge in the field of SNN. Currently, there are two mainstream methods, i.e., obtaining a converted SNN through converting a well-trained Artificial Neural Network (ANN) to its SNN counterpart or training an SNN directly. However, the inference time of a converted SNN is too long, while SNN training is generally very costly and inefficient. In this work, a new SNN training paradigm is proposed by combining the concepts of the two different training methods with the help of the pretrain technique and BP-based deep SNN training mechanism. We believe that the proposed paradigm is a more efficient pipeline for training SNNs. The pipeline includes pipeS for static data transfer tasks and pipeD for dynamic data transfer tasks. SOTA results are obtained in a large-scale event-driven dataset ES-ImageNet. For training acceleration, we achieve the same (or higher) best accuracy as similar LIF-SNNs using 1/10 training time on ImageNet-1K and 2/5 training time on ES-ImageNet and also provide a time-accuracy benchmark for a new dataset ES-UCF101. These experimental results reveal the similarity of the functions of parameters between ANNs and SNNs and also demonstrate the various potential applications of this SNN training pipeline.