论文标题
学习具有熵受训练的三元化(EC2T)的稀疏和三元神经网络
Learning Sparse & Ternary Neural Networks with Entropy-Constrained Trained Ternarization (EC2T)
论文作者
论文摘要
深度神经网络(DNN)在各种机器学习应用中表现出色。这些模型的容量(即参数数),赋予它们具有表现力的能力,并允许它们达到所需的性能。近年来,对将DNN部署到资源受限的设备(即移动设备)具有有限的能源,内存和计算预算的兴趣越来越大。为了解决这个问题,我们提出了熵受限的经过训练的三元化(EC2T),这是一个创建稀疏和三元神经网络的一般框架,这些框架在存储方面有效(例如,最多两个二进制掩码和两个完整的值都需要为了节省一个重量矩阵)和计算(例如,MAC操作都减少了两个乘以累积的乘以累积。这种方法包括两个步骤。首先,通过缩放预训练模型的尺寸(即其宽度和深度)来创建超级网络。随后,此超网络同时修剪(使用熵约束),并在训练过程中进行量化(即三元值分配),从而导致稀疏和三元网络表示。我们验证CIFAR-10,CIFAR-100和Imagenet数据集中提出的方法,显示其在图像分类任务中的有效性。
Deep neural networks (DNN) have shown remarkable success in a variety of machine learning applications. The capacity of these models (i.e., number of parameters), endows them with expressive power and allows them to reach the desired performance. In recent years, there is an increasing interest in deploying DNNs to resource-constrained devices (i.e., mobile devices) with limited energy, memory, and computational budget. To address this problem, we propose Entropy-Constrained Trained Ternarization (EC2T), a general framework to create sparse and ternary neural networks which are efficient in terms of storage (e.g., at most two binary-masks and two full-precision values are required to save a weight matrix) and computation (e.g., MAC operations are reduced to a few accumulations plus two multiplications). This approach consists of two steps. First, a super-network is created by scaling the dimensions of a pre-trained model (i.e., its width and depth). Subsequently, this super-network is simultaneously pruned (using an entropy constraint) and quantized (that is, ternary values are assigned layer-wise) in a training process, resulting in a sparse and ternary network representation. We validate the proposed approach in CIFAR-10, CIFAR-100, and ImageNet datasets, showing its effectiveness in image classification tasks.