论文标题
使用基于指标的解释(CSUME)选择多类ECG的核心设定选择
Core-set Selection Using Metrics-based Explanations (CSUME) for multiclass ECG
论文作者
论文摘要
基于深度学习的医疗保健决策支持系统,例如检测不规则的心律。大量医疗保健数据的收集和处理是一个连续的过程。渴望数据深度学习模型(DL)的性能高度取决于数据的数量和质量。虽然已经通过研究充分建立了对数据数量的需求,但我们展示了选择高质量数据的选择如何改善深度学习模型的性能。在这项工作中,我们将心电图(ECG)数据作为案例研究,并为算法开发人员提出了模型性能改进方法,该方法从多级ECG数据的传入流中选择了最有用的数据样本。我们的核心选择方法使用基于指标的解释来选择最有用的ECG数据样本。这也提供了一种理解(对于算法开发人员),即为什么选择样本比其他样本更有用,以改善深度学习模型的性能。我们的实验结果表明,精度为9.67%和8.69%,召回率提高,大幅度培训数据量减少了50%。此外,我们提出的方法宣称来自传入数据流的ECG样品的质量和注释。它允许自动检测单个数据样本,而这些样本不会导致模型学习,从而最大程度地减少对模型性能的负面影响。我们通过尝试不同的数据集和深度学习体系结构来进一步讨论方法的潜在普遍性。
The adoption of deep learning-based healthcare decision support systems such as the detection of irregular cardiac rhythm is hindered by challenges such as lack of access to quality data and the high costs associated with the collection and annotation of data. The collection and processing of large volumes of healthcare data is a continuous process. The performance of data-hungry Deep Learning models (DL) is highly dependent on the quantity and quality of the data. While the need for data quantity has been established through research adequately, we show how a selection of good quality data improves deep learning model performance. In this work, we take Electrocardiogram (ECG) data as a case study and propose a model performance improvement methodology for algorithm developers, that selects the most informative data samples from incoming streams of multi-class ECG data. Our Core-Set selection methodology uses metrics-based explanations to select the most informative ECG data samples. This also provides an understanding (for algorithm developers) as to why a sample was selected as more informative over others for the improvement of deep learning model performance. Our experimental results show a 9.67% and 8.69% precision and recall improvement with a significant training data volume reduction of 50%. Additionally, our proposed methodology asserts the quality and annotation of ECG samples from incoming data streams. It allows automatic detection of individual data samples that do not contribute to model learning thus minimizing possible negative effects on model performance. We further discuss the potential generalizability of our approach by experimenting with a different dataset and deep learning architecture.