论文标题

柯尔特:循环重叠的彩票,以更快的卷积神经网络修剪

COLT: Cyclic Overlapping Lottery Tickets for Faster Pruning of Convolutional Neural Networks

论文作者

Hossain, Md. Ismail, Rakib, Mohammed, Elahi, M. M. Lutfe, Mohammed, Nabeel, Rahman, Shafin

论文摘要

修剪是指从神经网络中消除琐碎的权重。修剪后产生的过度参数化模型中的子网络通常称为彩票。这项研究旨在从一组彩票中产生获胜的彩票,这些彩票可以达到与原始未经修复的网络相似的准确性。我们通过从头开始的数据分裂和循环再培训介绍了一张新颖的获胜票,名为“环状重叠彩票(COLT)”。我们应用了一种环状修剪算法,该算法仅保留在不同数据段训练的不同修剪模型的重叠权重。我们的结果表明,柯尔特可以在保持较高的稀有度的同时获得类似的精度(通过未经修复的模型获得)。我们表明,柯尔特的准确性与彩票假设(LTH)的获胜门票相当,有时会更好。此外,可以使用比流行的迭代幅度修剪(IMP)方法生成的门票更少的迭代来生成小马队。此外,我们还注意到在大型数据集上生成的柯尔特可以将其转移到小型数据集上,而不会损害性能,从而证明了其概括能力。我们在CIFAR-10,CIFAR-100和TINYIMAGENET数据集上进行了所有实验,并报告了比最先进的方法更高的性能。

Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within an overparameterized model produced after pruning are often called Lottery tickets. This research aims to generate winning lottery tickets from a set of lottery tickets that can achieve similar accuracy to the original unpruned network. We introduce a novel winning ticket called Cyclic Overlapping Lottery Ticket (COLT) by data splitting and cyclic retraining of the pruned network from scratch. We apply a cyclic pruning algorithm that keeps only the overlapping weights of different pruned models trained on different data segments. Our results demonstrate that COLT can achieve similar accuracies (obtained by the unpruned model) while maintaining high sparsities. We show that the accuracy of COLT is on par with the winning tickets of Lottery Ticket Hypothesis (LTH) and, at times, is better. Moreover, COLTs can be generated using fewer iterations than tickets generated by the popular Iterative Magnitude Pruning (IMP) method. In addition, we also notice COLTs generated on large datasets can be transferred to small ones without compromising performance, demonstrating its generalizing capability. We conduct all our experiments on Cifar-10, Cifar-100 & TinyImageNet datasets and report superior performance than the state-of-the-art methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源