论文标题

解释方法解释深度学习的植物压力表型的有用性

Usefulness of interpretability methods to explain deep learning based plant stress phenotyping

论文作者

Nagasubramanian, Koushik, Singh, Asheesh K., Singh, Arti, Sarkar, Soumik, Ganapathysubramanian, Baskar

论文摘要

深度学习技术已成功部署以自动化植物压力识别和定量。近年来,越来越多的推动培训模型的推动力-I.E。通过视觉突出图像特征对于分类决策至关重要,这证明了他们的分类决策是合理的。期望的是,受过训练的网络模型利用了模仿植物病理学家使用的图像特征。在这项工作中,我们比较了一些最流行的可解释性方法:显着图,平滑果,带有指导性的背部,深层泰勒分解,集成梯度,层面相关性的传播和梯度时间输入,以解释深度学习模型。我们训练一个Densenet-121网络,以分类八种不同的大豆应力(生物和非生物)。使用在受控条件下捕获的健康和压力大豆传单的16,573 RGB图像的数据集,我们获得了95.05 \%的总体分类精度。对于测试数据的各种子集,我们将重要特征与人类专家确定的特征进行了比较。我们观察到,大多数可解释性方法将叶子的感染区域识别为某些(但不是全部)正确分类图像的重要特征。对于某些图像,可解释性方法的输出表明伪造特征相关性可能已用于正确对其进行分类。尽管这些可解释性方法的输出解释图可能彼此不同,但我们提倡将这些可解释性方法用作“假设产生”机制,这些机制可以推动科学见解。

Deep learning techniques have been successfully deployed for automating plant stress identification and quantification. In recent years, there is a growing push towards training models that are interpretable -i.e. that justify their classification decisions by visually highlighting image features that were crucial for classification decisions. The expectation is that trained network models utilize image features that mimic visual cues used by plant pathologists. In this work, we compare some of the most popular interpretability methods: Saliency Maps, SmoothGrad, Guided Backpropogation, Deep Taylor Decomposition, Integrated Gradients, Layer-wise Relevance Propagation and Gradient times Input, for interpreting the deep learning model. We train a DenseNet-121 network for the classification of eight different soybean stresses (biotic and abiotic). Using a dataset consisting of 16,573 RGB images of healthy and stressed soybean leaflets captured under controlled conditions, we obtained an overall classification accuracy of 95.05 \%. For a diverse subset of the test data, we compared the important features with those identified by a human expert. We observed that most interpretability methods identify the infected regions of the leaf as important features for some -- but not all -- of the correctly classified images. For some images, the output of the interpretability methods indicated that spurious feature correlations may have been used to correctly classify them. Although the output explanation maps of these interpretability methods may be different from each other for a given image, we advocate the use of these interpretability methods as `hypothesis generation' mechanisms that can drive scientific insight.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源