论文标题
探索二进制和尖峰神经网络之间的联系
Exploring the Connection Between Binary and Spiking Neural Networks
论文作者
论文摘要
片上边缘智能需要探索算法技术,以减少当前机器学习框架的计算要求。这项工作旨在弥合训练二进制神经网络和尖峰神经网络的最新算法进展,这两者都是由相同的动机驱动的,但两者之间的协同作用尚未得到充分探索。我们表明,在极端量化制度中的训练尖峰神经网络在大规模数据集(如Cifar- $ 100 $ and Imagenet)上几乎完全精确地精确。这项工作的一个重要含义是,可以通过迎合二进制神经网络的“内存”硬件加速器来启用二进制尖峰神经网络,而不会因二进制而遭受任何准确的降解。我们利用标准培训技术来针对非加速网络来通过转换过程来产生我们的尖峰网络,并进行广泛的经验分析,并探索简单的设计时间和运行时优化技术,以通过先前工作的级级来降低尖峰网络(用于二进制和完整计算模型)的推断网络的推理潜伏期(二进制和完整精确模型)。
On-chip edge intelligence has necessitated the exploration of algorithmic techniques to reduce the compute requirements of current machine learning frameworks. This work aims to bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks - both of which are driven by the same motivation and yet synergies between the two have not been fully explored. We show that training Spiking Neural Networks in the extreme quantization regime results in near full precision accuracies on large-scale datasets like CIFAR-$100$ and ImageNet. An important implication of this work is that Binary Spiking Neural Networks can be enabled by "In-Memory" hardware accelerators catered for Binary Neural Networks without suffering any accuracy degradation due to binarization. We utilize standard training techniques for non-spiking networks to generate our spiking networks by conversion process and also perform an extensive empirical analysis and explore simple design-time and run-time optimization techniques for reducing inference latency of spiking networks (both for binary and full-precision models) by an order of magnitude over prior work.