Zhao Benhan, Kang Xilin, Zhou Hao, Shi Ziyang, Li Lin, Zhou Guoxiong, Wan Fangying, Zhu Jiangzhang, Yan Yongming, Li Leheng, Wu Yulong
School of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China.
School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China.
Plants (Basel). 2025 Aug 24;14(17):2634. doi: 10.3390/plants14172634.
Plant disease segmentation has achieved significant progress with the help of artificial intelligence. However, deploying high-accuracy segmentation models in resource-limited settings faces three key challenges, as follows: (A) Traditional dense attention mechanisms incur quadratic computational complexity growth (O(n2d)), rendering them ill-suited for low-power hardware. (B) Naturally sparse spatial distributions and large-scale variations in the lesions on leaves necessitate models that concurrently capture long-range dependencies and local details. (C) Complex backgrounds and variable lighting in field images often induce segmentation errors. To address these challenges, we propose Sparse-MoE-SAM, an efficient framework based on an enhanced Segment Anything Model (SAM). This deep learning framework integrates sparse attention mechanisms with a two-stage mixture of experts (MoE) decoder. The sparse attention dynamically activates key channels aligned with lesion sparsity patterns, reducing self-attention complexity while preserving long-range context. Stage 1 of the MoE decoder performs coarse-grained boundary localization; Stage 2 achieves fine-grained segmentation by leveraging specialized experts within the MoE, significantly enhancing edge discrimination accuracy. The expert repository-comprising standard convolutions, dilated convolutions, and depthwise separable convolutions-dynamically routes features through optimized processing paths based on input texture and lesion morphology. This enables robust segmentation across diverse leaf textures and plant developmental stages. Further, we design a sparse attention-enhanced Atrous Spatial Pyramid Pooling (ASPP) module to capture multi-scale contexts for both extensive lesions and small spots. Evaluations on three heterogeneous datasets (PlantVillage Extended, CVPPP, and our self-collected field images) show that Sparse-MoE-SAM achieves a mean Intersection-over-Union (mIoU) of 94.2%-surpassing standard SAM by 2.5 percentage points-while reducing computational costs by 23.7% compared to the original SAM baseline. The model also demonstrates balanced performance across disease classes and enhanced hardware compatibility. Our work validates that integrating sparse attention with MoE mechanisms sustains accuracy while drastically lowering computational demands, enabling the scalable deployment of plant disease segmentation models on mobile and edge devices.
借助人工智能,植物病害分割已取得显著进展。然而,在资源有限的环境中部署高精度分割模型面临三个关键挑战,具体如下:(A)传统的密集注意力机制会导致二次计算复杂度增长(O(n2d)),使其不适用于低功耗硬件。(B)叶片上病斑的自然稀疏空间分布和大规模变化需要模型同时捕捉长距离依赖关系和局部细节。(C)田间图像中复杂的背景和多变的光照常常会引发分割错误。为应对这些挑战,我们提出了Sparse-MoE-SAM,这是一个基于增强型分割一切模型(SAM)的高效框架。这个深度学习框架将稀疏注意力机制与两阶段专家混合(MoE)解码器相结合。稀疏注意力动态激活与病斑稀疏模式对齐的关键通道,在保留长距离上下文的同时降低自注意力复杂度。MoE解码器的第一阶段执行粗粒度边界定位;第二阶段通过利用MoE中的专门专家实现细粒度分割,显著提高边缘辨别精度。由标准卷积、空洞卷积和深度可分离卷积组成的专家库根据输入纹理和病斑形态,通过优化的处理路径动态路由特征。这使得能够在不同的叶片纹理和植物发育阶段进行稳健分割。此外,我们设计了一个稀疏注意力增强的空洞空间金字塔池化(ASPP)模块,以捕捉大面积病斑和小斑点的多尺度上下文。在三个异构数据集(PlantVillage Extended、CVPPP和我们自己收集的田间图像)上的评估表明,Sparse-MoE-SAM实现了94.2%的平均交并比(mIoU)——比标准SAM高出2.5个百分点——同时与原始SAM基线相比,计算成本降低了23.7%。该模型还在不同病害类别上表现出平衡的性能,并增强了硬件兼容性。我们的工作验证了将稀疏注意力与MoE机制相结合在大幅降低计算需求的同时保持了准确性,从而能够在移动和边缘设备上可扩展地部署植物病害分割模型。