论文标题
MLPERF移动推理基准测试
MLPerf Mobile Inference Benchmark
论文作者
论文摘要
本文介绍了第一个行业标准的开源机器学习(ML)基准,以允许对具有不同AI芯片和软件堆栈的移动设备进行perfor perfor和精确评估。该基准借鉴了领先的移动SOC供应商,ML-Framework提供商和模型生产商的专业知识。它包括一套套件,这些模型与标准数据集,质量指标和运行规则一起运行。我们描述了该特定领域的ML基准的设计和实现。当前的基准版本是用于不同计算机视觉和自然语言处理任务的移动应用程序。该基准还支持非智能手机设备,例如笔记本电脑和移动PC。前两轮的基准结果揭示了基础移动ML系统堆栈的压倒性复杂性,这强调了移动ML性能分析中透明度的需求。结果还表明,通过ML堆栈取得的步伐提高了性能。在六个月内,离线吞吐量提高了3倍,而潜伏期降低了12倍。 ML是一个不断发展的领域,其用例,模型,数据集和质量目标变化。 MLPERF Mobile将发展并作为开源社区框架,以指导移动AI的研究和创新。
This paper presents the first industry-standard open-source machine learning (ML) benchmark to allow perfor mance and accuracy evaluation of mobile devices with different AI chips and software stacks. The benchmark draws from the expertise of leading mobile-SoC vendors, ML-framework providers, and model producers. It comprises a suite of models that operate with standard data sets, quality metrics and run rules. We describe the design and implementation of this domain-specific ML benchmark. The current benchmark version comes as a mobile app for different computer vision and natural language processing tasks. The benchmark also supports non-smartphone devices, such as laptops and mobile PCs. Benchmark results from the first two rounds reveal the overwhelming complexity of the underlying mobile ML system stack, emphasizing the need for transparency in mobile ML performance analysis. The results also show that the strides being made all through the ML stack improve performance. Within six months, offline throughput improved by 3x, while latency reduced by as much as 12x. ML is an evolving field with changing use cases, models, data sets and quality targets. MLPerf Mobile will evolve and serve as an open-source community framework to guide research and innovation for mobile AI.