论文标题

UCP:深度卷积神经网络压缩和加速度的均匀通道修剪

UCP: Uniform Channel Pruning for Deep Convolutional Neural Networks Compression and Acceleration

论文作者

Chang, Jingfei, Lu, Yang, Xue, Ping, Wei, Xing, Wei, Zhen

论文摘要

为了将深CNN应用于移动终端和便携式设备,许多学者最近致力于压缩和加速深层卷积神经网络。基于此,我们提出了一种新颖的均匀通道修复(UCP)方法来修剪深CNN,并使用改良的挤压和兴奋块(MSEB)来衡量卷积层中通道的重要性。直接修剪了不重要的通道,包括与它们相关的卷积内核,这大大降低了存储成本和计算数量。 Resnet中有两种类型的残留块。对于带瓶颈的重新连接,我们使用带有传统CNN的修剪方法来修剪块中间的3x3卷积层。对于具有基本残留块的重新连接,我们提出了一种在同一阶段始终修剪所有残留块的方法,以确保紧凑的网络结构在尺寸上是正确的。考虑到该网络在修剪后会丢失大量信息,并且修剪幅度的幅度越大,丢失的信息越多,我们不会选择微调,而是从头开始重新调整以恢复修剪后网络的准确性。最后,我们在CIFAR-10,CIFAR-100和ILSVRC-2012上验证了我们的方法进行图像分类。结果表明,在修剪率很小时,紧凑型网络在从头开始重新划痕后的性能优于原始网络。即使修剪幅度很大,也可以稍微稍微降低准确性。在CIFAR-100上,当分别将参数和62%的参数降低到82%和62%时,VGG-19的准确性甚至在再培训后甚至提高了0.54%。

To apply deep CNNs to mobile terminals and portable devices, many scholars have recently worked on the compressing and accelerating deep convolutional neural networks. Based on this, we propose a novel uniform channel pruning (UCP) method to prune deep CNN, and the modified squeeze-and-excitation blocks (MSEB) is used to measure the importance of the channels in the convolutional layers. The unimportant channels, including convolutional kernels related to them, are pruned directly, which greatly reduces the storage cost and the number of calculations. There are two types of residual blocks in ResNet. For ResNet with bottlenecks, we use the pruning method with traditional CNN to trim the 3x3 convolutional layer in the middle of the blocks. For ResNet with basic residual blocks, we propose an approach to consistently prune all residual blocks in the same stage to ensure that the compact network structure is dimensionally correct. Considering that the network loses considerable information after pruning and that the larger the pruning amplitude is, the more information that will be lost, we do not choose fine-tuning but retrain from scratch to restore the accuracy of the network after pruning. Finally, we verified our method on CIFAR-10, CIFAR-100 and ILSVRC-2012 for image classification. The results indicate that the performance of the compact network after retraining from scratch, when the pruning rate is small, is better than the original network. Even when the pruning amplitude is large, the accuracy can be maintained or decreased slightly. On the CIFAR-100, when reducing the parameters and FLOPs up to 82% and 62% respectively, the accuracy of VGG-19 even improved by 0.54% after retraining.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源