论文标题

将Rocket执行效率任务:使用LightWaves的多元时间序列分类

Taking ROCKET on an Efficiency Mission: Multivariate Time Series Classification with LightWaveS

论文作者

Pantiskas, Leonardos, Verstoep, Kees, Hoogendoorn, Mark, Bal, Henri

论文摘要

如今,随着医疗保健和行业等领域的传感器数量的增加,多元时间序列分类(MTSC)的问题变得越来越相关,并且是机器和深度学习方法的主要目标。它们在实际环境中的扩展采用导致重点从追求更高的预测准确性,复杂模型转向可实用的可部署解决方案,这些解决方案平衡了准确性和参数,例如预测速度。最近引起关注的MTSC模型是基于随机卷积内核的火箭弹,这既是其非常快速的训练过程及其最先进的准确性。但是,它使用的大量功能可能对推理时间有害。检查其理论背景和局限性使我们能够解决潜在的缺点和当前的光线:准确的MTSC框架,该框架在训练和推理过程中既快速又快。具体来说,利用小波散射转换和分布式特征选择,我们设法创建了一种仅采用2.5%的火箭功能的解决方案,同时实现了与最近的MTSC模型相当的精度。 LightWaves在多个计算节点以及训练过程中输入通道的数量也很好地缩放。此外,它可以通过仅保留最有用的渠道来大大减少输入大小,并为MTSC问题提供洞察力。我们介绍了三个版本的算法及其在分布式训练时间和可扩展性,准确性和推理速度上的结果。我们表明,与在边缘设备上推断在数据集上的推断相比,与火箭相比,我们实现了从9倍到53倍的速度。

Nowadays, with the rising number of sensors in sectors such as healthcare and industry, the problem of multivariate time series classification (MTSC) is getting increasingly relevant and is a prime target for machine and deep learning approaches. Their expanding adoption in real-world environments is causing a shift in focus from the pursuit of ever-higher prediction accuracy with complex models towards practical, deployable solutions that balance accuracy and parameters such as prediction speed. An MTSC model that has attracted attention recently is ROCKET, based on random convolutional kernels, both because of its very fast training process and its state-of-the-art accuracy. However, the large number of features it utilizes may be detrimental to inference time. Examining its theoretical background and limitations enables us to address potential drawbacks and present LightWaveS: a framework for accurate MTSC, which is fast both during training and inference. Specifically, utilizing wavelet scattering transformation and distributed feature selection, we manage to create a solution that employs just 2.5% of the ROCKET features, while achieving accuracy comparable to recent MTSC models. LightWaveS also scales well across multiple compute nodes and with the number of input channels during training. In addition, it can significantly reduce the input size and provide insight to an MTSC problem by keeping only the most useful channels. We present three versions of our algorithm and their results on distributed training time and scalability, accuracy, and inference speedup. We show that we achieve speedup ranging from 9x to 53x compared to ROCKET during inference on an edge device, on datasets with comparable accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源