论文标题

深度部分更新:朝着沟通有效更新以进行设备推理

Deep Partial Updating: Towards Communication Efficient Updating for On-device Inference

论文作者

Qu, Zhongnan, Liu, Cong, Thiele, Lothar

论文摘要

新兴的边缘情报应用程序要求服务器重新训练和更新部署在远程边缘节点上的深神经网络,以利用新收集的数据示例。不幸的是,由于高度严格的通信资源,在实践中可能不可能不断向这些边缘节点发送全面更新的权重。在本文中,我们提出了重量的深层部分更新范式,该范式巧妙地选择了一小部分的权重以在每个服务器到边缘通信回合中进行更新,同时与完整更新相比实现了相似的性能。我们的方法是通过分析上限的部分更新和完整更新之间的损失差异来建立的,并且只能更新权重,从而对上限产生最大的贡献。广泛的实验结果证明了我们部分更新方法的功效,该方法在更新少量的重量时达到了高推理精度。

Emerging edge intelligence applications require the server to retrain and update deep neural networks deployed on remote edge nodes to leverage newly collected data samples. Unfortunately, it may be impossible in practice to continuously send fully updated weights to these edge nodes due to the highly constrained communication resource. In this paper, we propose the weight-wise deep partial updating paradigm, which smartly selects a small subset of weights to update in each server-to-edge communication round, while achieving a similar performance compared to full updating. Our method is established through analytically upper-bounding the loss difference between partial updating and full updating, and only updates the weights which make the largest contributions to the upper bound. Extensive experimental results demonstrate the efficacy of our partial updating methodology which achieves a high inference accuracy while updating a rather small number of weights.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源