Wöber Wilfried, Mehnen Lars, Sykacek Peter, Meimberg Harald
Department of Integrative Biology and Biodiversity Research, Institute of Integrative Conservation Research, University of Natural Resources and Life Sciences, Gregor Mendel Str. 33, 1080 Vienna, Austria.
Department Industrial Engineering, University of Applied Sciences Technikum Wien, Höchstädtplatz 6, 1200 Vienna, Austria.
Plants (Basel). 2021 Dec 5;10(12):2674. doi: 10.3390/plants10122674.
Recent progress in machine learning and deep learning has enabled the implementation of plant and crop detection using systematic inspection of the leaf shapes and other morphological characters for identification systems for precision farming. However, the models used for this approach tend to become black-box models, in the sense that it is difficult to trace characters that are the base for the classification. The interpretability is therefore limited and the explanatory factors may not be based on reasonable visible characters. We investigate the explanatory factors of recent machine learning and deep learning models for plant classification tasks. Based on a Daucus carota and a Beta vulgaris image data set, we implement plant classification models and compare those models by their predictive performance as well as explainability. For comparison we implemented a feed forward convolutional neuronal network as a default model. To evaluate the performance, we trained an unsupervised Bayesian Gaussian process latent variable model as well as a convolutional autoencoder for feature extraction and rely on a support vector machine for classification. The explanatory factors of all models were extracted and analyzed. The experiments show, that feed forward convolutional neuronal networks (98.24% and 96.10% mean accuracy) outperforms the Bayesian Gaussian process latent variable pipeline (92.08% and 94.31% mean accuracy) as well as the convolutional autoenceoder pipeline (92.38% and 93.28% mean accuracy) based approaches in terms of classification accuracy, even though not significant for Beta vulgaris images. Additionally, we found that the neuronal network used biological uninterpretable image regions for the plant classification task. In contrast to that, the unsupervised learning models rely on explainable visual characters. We conclude that supervised convolutional neuronal networks must be used carefully to ensure biological interpretability. We recommend unsupervised machine learning, careful feature investigation, and statistical feature analysis for biological applications.
机器学习和深度学习的最新进展使得通过系统检查叶片形状和其他形态特征来实现植物和作物检测成为可能,这适用于精准农业的识别系统。然而,用于这种方法的模型往往会变成黑箱模型,也就是说,很难追踪作为分类基础的特征。因此,其可解释性有限,且解释因素可能并非基于合理的可见特征。我们研究了用于植物分类任务的最新机器学习和深度学习模型的解释因素。基于胡萝卜和甜菜的图像数据集,我们实现了植物分类模型,并通过预测性能和可解释性对这些模型进行比较。为了进行比较,我们实现了一个前馈卷积神经网络作为默认模型。为了评估性能,我们训练了一个无监督贝叶斯高斯过程潜变量模型以及一个用于特征提取的卷积自动编码器,并依靠支持向量机进行分类。提取并分析了所有模型的解释因素。实验表明,在前馈卷积神经网络(平均准确率分别为98.24%和96.10%)在分类准确率方面优于基于贝叶斯高斯过程潜变量管道(平均准确率分别为92.08%和94.31%)以及卷积自动编码器管道(平均准确率分别为92.38%和93.28%)的方法,尽管对于甜菜图像来说差异不显著。此外,我们发现神经网络在植物分类任务中使用了生物学上不可解释的图像区域。相比之下,无监督学习模型依赖于可解释的视觉特征。我们得出结论,必须谨慎使用监督卷积神经网络以确保生物学可解释性。我们建议在生物学应用中使用无监督机器学习、仔细进行特征研究和统计特征分析。