• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种结合卷积和视觉Transformer的双分支模型用于作物病害分类。

A dual-branch model combining convolution and vision transformer for crop disease classification.

作者信息

Meng Qingduan, Guo Jiadong, Zhang Hui, Zhou Yaoqi, Zhang Xiaoling

机构信息

College of Information Engineering, Henan University of Science and Technology, Luoyang, Henan, China.

出版信息

PLoS One. 2025 Apr 24;20(4):e0321753. doi: 10.1371/journal.pone.0321753. eCollection 2025.

DOI:10.1371/journal.pone.0321753
PMID:40273041
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12021192/
Abstract

Computer vision holds tremendous potential in crop disease classification, but the complex texture and shape characteristics of crop diseases make disease classification challenging. To address these issues, this paper proposes a dual-branch model for crop disease classification, which combines Convolutional Neural Network (CNN) with Vision Transformer (ViT). Here, the convolutional branch is utilized to capture the local features while the Transformer branch is utilized to handle global features. A learnable parameter is used to achieve a linear weighted fusion of these two types of features. An Aggregated Local Perceptive Feed Forward Layer (ALP-FFN) is introduced to enhance the model's representation capability by introducing locality into the Transformer encoder. Furthermore, this paper constructs a lightweight Transformer block using ALP-FFN and a linear self-attention mechanism to reduce the model's parameters and computational cost. The proposed model achieves an exceptional classification accuracy of 99.71% on the PlantVillage dataset with only 4.9M parameters and 0.62G FLOPs, surpassing the state-of-the-art TNT-S model (accuracy: 99.11%, parameters: 23.31M, FLOPs: 4.85G) by 0.6%. On the Potato Leaf dataset, the model attains 98.78% classification accuracy, outperforming the advanced ResNet-18 model (accuracy: 98.05%, parameters: 11.18M, FLOPs: 1.82G) by 0.73%. The model proposed in this paper effectively combines the advantages of CNN and ViT while maintaining a lightweight design, providing an effective method for the precise identification of crop diseases.

摘要

计算机视觉在作物病害分类中具有巨大潜力,但作物病害复杂的纹理和形状特征使病害分类具有挑战性。为解决这些问题,本文提出一种用于作物病害分类的双分支模型,该模型将卷积神经网络(CNN)与视觉Transformer(ViT)相结合。在此,卷积分支用于捕捉局部特征,而Transformer分支用于处理全局特征。使用一个可学习参数来实现这两种特征的线性加权融合。引入聚合局部感知前馈层(ALP-FFN),通过将局部性引入Transformer编码器来增强模型的表示能力。此外,本文使用ALP-FFN和线性自注意力机制构建了一个轻量级的Transformer模块,以减少模型的参数和计算成本。所提出的模型在PlantVillage数据集上仅用490万个参数和0.62G FLOPs就达到了99.71%的卓越分类准确率,比当前最优的TNT-S模型(准确率:99.11%,参数:2331万个,FLOPs:4.85G)高出0.6%。在马铃薯叶片数据集上,该模型达到了98.78%的分类准确率,比先进的ResNet-18模型(准确率:98.05%,参数:1118万个,FLOPs:1.82G)高出0.73%。本文提出的模型有效结合了CNN和ViT的优点,同时保持了轻量级设计,为作物病害的精确识别提供了一种有效方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/288877d33cc7/pone.0321753.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/0fbbec1b4907/pone.0321753.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/1055fce79f7a/pone.0321753.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/867c142ab6ac/pone.0321753.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/5cc350c5e244/pone.0321753.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/0ebbc6f4982b/pone.0321753.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/5100d1b532aa/pone.0321753.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/ad0a40a663bd/pone.0321753.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/432bf15f0f41/pone.0321753.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/23e955141032/pone.0321753.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/bf65eaa98451/pone.0321753.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/47073a0875c6/pone.0321753.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/c2c535dfa61b/pone.0321753.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/288877d33cc7/pone.0321753.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/0fbbec1b4907/pone.0321753.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/1055fce79f7a/pone.0321753.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/867c142ab6ac/pone.0321753.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/5cc350c5e244/pone.0321753.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/0ebbc6f4982b/pone.0321753.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/5100d1b532aa/pone.0321753.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/ad0a40a663bd/pone.0321753.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/432bf15f0f41/pone.0321753.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/23e955141032/pone.0321753.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/bf65eaa98451/pone.0321753.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/47073a0875c6/pone.0321753.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/c2c535dfa61b/pone.0321753.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/174c/12021192/288877d33cc7/pone.0321753.g013.jpg

相似文献

1
A dual-branch model combining convolution and vision transformer for crop disease classification.一种结合卷积和视觉Transformer的双分支模型用于作物病害分类。
PLoS One. 2025 Apr 24;20(4):e0321753. doi: 10.1371/journal.pone.0321753. eCollection 2025.
2
Crop Disease Identification by Fusing Multiscale Convolution and Vision Transformer.基于多尺度卷积和视觉Transformer 的作物病害识别
Sensors (Basel). 2023 Jun 29;23(13):6015. doi: 10.3390/s23136015.
3
Enhanced Pneumonia Detection in Chest X-Rays Using Hybrid Convolutional and Vision Transformer Networks.使用混合卷积和视觉Transformer网络增强胸部X光片中的肺炎检测
Curr Med Imaging. 2025;21:e15734056326685. doi: 10.2174/0115734056326685250101113959.
4
EfficientRMT-Net-An Efficient ResNet-50 and Vision Transformers Approach for Classifying Potato Plant Leaf Diseases.高效 RMT-Net:一种基于 ResNet-50 和 Vision Transformers 的马铃薯叶片病害分类方法。
Sensors (Basel). 2023 Nov 30;23(23):9516. doi: 10.3390/s23239516.
5
Efficient agricultural pest classification using vision transformer with hybrid pooled multihead attention.利用融合池多头注意力的视觉转换器实现高效农业虫害分类。
Comput Biol Med. 2024 Jul;177:108584. doi: 10.1016/j.compbiomed.2024.108584. Epub 2024 May 13.
6
Plant-CNN-ViT: Plant Classification with Ensemble of Convolutional Neural Networks and Vision Transformer.植物-CNN-ViT:基于卷积神经网络与视觉Transformer集成的植物分类方法
Plants (Basel). 2023 Jul 14;12(14):2642. doi: 10.3390/plants12142642.
7
Interactively Fusing Global and Local Features for Benign and Malignant Classification of Breast Ultrasound Images.交互式融合全局和局部特征用于乳腺超声图像的良恶性分类
Ultrasound Med Biol. 2025 Mar;51(3):525-534. doi: 10.1016/j.ultrasmedbio.2024.11.014. Epub 2024 Dec 20.
8
ScribFormer: Transformer Makes CNN Work Better for Scribble-Based Medical Image Segmentation.ScribFormer:Transformer 使 CNN 更适用于基于草图的医学图像分割。
IEEE Trans Med Imaging. 2024 Jun;43(6):2254-2265. doi: 10.1109/TMI.2024.3363190. Epub 2024 Jun 3.
9
PMVT: a lightweight vision transformer for plant disease identification on mobile devices.PMVT:一种用于移动设备上植物病害识别的轻量级视觉变换器。
Front Plant Sci. 2023 Sep 26;14:1256773. doi: 10.3389/fpls.2023.1256773. eCollection 2023.
10
Dual encoder network with transformer-CNN for multi-organ segmentation.基于 Transformer-CNN 的双编码器网络的多器官分割。
Med Biol Eng Comput. 2023 Mar;61(3):661-671. doi: 10.1007/s11517-022-02723-9. Epub 2022 Dec 29.

本文引用的文献

1
PSLT: A Light-Weight Vision Transformer With Ladder Self-Attention and Progressive Shift.PSLT:一种具有阶梯式自注意力和渐进式移位的轻量级视觉Transformer
IEEE Trans Pattern Anal Mach Intell. 2023 Sep;45(9):11120-11135. doi: 10.1109/TPAMI.2023.3265499. Epub 2023 Aug 7.
2
Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification.基于深度神经网络的植物病害叶片图像分类识别
Comput Intell Neurosci. 2016;2016:3289801. doi: 10.1155/2016/3289801. Epub 2016 Jun 22.