论文标题

FTBNN:重新考虑1位CNN的非线性并超越

FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond

论文作者

Su, Zhuo, Fang, Linpu, Guo, Deke, Hu, Dewen, Pietikäinen, Matti, Liu, Li

论文摘要

二进制神经网络(BNN),重量和激活都被分为1位,由于其高度加速计算的巨大好处,并大大减少了对资源约束设备开发的开发,因此近年来已经对其进行了广泛的研究。与以前倾向于减少训练BNN结构的量化误差的方法相反,我们认为二进制卷积过程具有越来越多的线性性,该目标对最小化误差的目标越来越多,这反过来又阻碍了BNN的歧视能力。在本文中,我们重新调查并调整适当的非线性模块来解决该矛盾,从而使基线强大,从而在准确性和训练效率方面实现了大规模图像网数据集的最新性能。再进一步,我们发现所提出的BNN模型仍然具有很大的潜力,可以通过更好地利用有效的二进制操作而不会丢失准确性来压缩。此外,BNN模型的有限容量也可以在小组执行的帮助下增加。基于这些见解,即使计算成本较小,我们也能以额外的4〜5%的TOP-1准确性增长来提高基线。我们的代码将在https://github.com/zhuogege1943/ftbnn上公开。

Binary neural networks (BNNs), where both weights and activations are binarized into 1 bit, have been widely studied in recent years due to its great benefit of highly accelerated computation and substantially reduced memory footprint that appeal to the development of resource constrained devices. In contrast to previous methods tending to reduce the quantization error for training BNN structures, we argue that the binarized convolution process owns an increasing linearity towards the target of minimizing such error, which in turn hampers BNN's discriminative ability. In this paper, we re-investigate and tune proper non-linear modules to fix that contradiction, leading to a strong baseline which achieves state-of-the-art performance on the large-scale ImageNet dataset in terms of accuracy and training efficiency. To go further, we find that the proposed BNN model still has much potential to be compressed by making a better use of the efficient binary operations, without losing accuracy. In addition, the limited capacity of the BNN model can also be increased with the help of group execution. Based on these insights, we are able to improve the baseline with an additional 4~5% top-1 accuracy gain even with less computational cost. Our code will be made public at https://github.com/zhuogege1943/ftbnn.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源