• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CFI-Former:通过多粒度感知查询注意力变换器实现高效车道检测

CFI-Former: Efficient lane detection by multi-granularity perceptual query attention transformer.

作者信息

Gao Rong, Hu Siqi, Yan Lingyu, Zhang Lefei, Wu Jia

机构信息

School of Computer Science, Hubei University of Technology, Wuhan, 430068, China; State Key Laboratory for Novel Software Technology at Nanjing University, Nanjing, 210023, China.

School of Computer Science, Hubei University of Technology, Wuhan, 430068, China.

出版信息

Neural Netw. 2025 Jul;187:107347. doi: 10.1016/j.neunet.2025.107347. Epub 2025 Mar 15.

DOI:10.1016/j.neunet.2025.107347
PMID:40101557
Abstract

Benefiting from the booming development of Transformer methods, the performance of lane detection tasks has been rapidly improved. However, due to the influence of inaccurate lane line shape constraints, the query sequences of existing transformer-based lane line detection methods contain a large number of repetitive and invalid information regions, which leads to redundant information in the detection region and makes the processing of information on localized feature details of the lanes biased. In this paper, a multi-granularity perceptual query attention transformer lane detection method, CFI-Former, is proposed to achieve more accurate lane detection. Specifically, a multi-granularity perceptual query attention (GQA) module is designed to extract lane local detail information. By a two-stage query from coarse to fine, redundant key-value pairs with low information relevance are first filtered out, and then fine-grained token-to-token attention is executed on the remaining candidate regions. This module emphasizes the multi-granularity nuances of lane features from global to local, leading to more effective models based on lane line shape constraints. In addition, weighted adaptive LIoU loss (L) is proposed to improve lane detection in more challenging scenarios by adaptively increasing the relative gradient of high IoU lane objects and the weight of the loss. Extensive experiments show that CFI-Former outperforms the baseline on two popular lane detection benchmark datasets.

摘要

受益于Transformer方法的蓬勃发展,车道检测任务的性能得到了快速提升。然而,由于车道线形状约束不准确的影响,现有的基于Transformer的车道线检测方法的查询序列包含大量重复和无效的信息区域,这导致检测区域中存在冗余信息,并使得对车道局部特征细节信息的处理产生偏差。本文提出了一种多粒度感知查询注意力Transformer车道检测方法CFI-Former,以实现更精确的车道检测。具体来说,设计了一个多粒度感知查询注意力(GQA)模块来提取车道局部细节信息。通过从粗到细的两阶段查询,首先过滤掉信息相关性低的冗余键值对,然后对剩余的候选区域执行细粒度的token-to-token注意力。该模块强调了车道特征从全局到局部的多粒度细微差别,从而基于车道线形状约束得到更有效的模型。此外,还提出了加权自适应LIoU损失(L),通过自适应增加高IoU车道对象的相对梯度和损失权重,在更具挑战性的场景中改进车道检测。大量实验表明,CFI-Former在两个流行的车道检测基准数据集上优于基线。

相似文献

1
CFI-Former: Efficient lane detection by multi-granularity perceptual query attention transformer.CFI-Former:通过多粒度感知查询注意力变换器实现高效车道检测
Neural Netw. 2025 Jul;187:107347. doi: 10.1016/j.neunet.2025.107347. Epub 2025 Mar 15.
2
DiagSWin: A multi-scale vision transformer with diagonal-shaped windows for object detection and segmentation.DiagSWin:一种具有对角线形状窗口的多尺度视觉转换器,用于目标检测和分割。
Neural Netw. 2024 Dec;180:106653. doi: 10.1016/j.neunet.2024.106653. Epub 2024 Aug 22.
3
Transformer guided self-adaptive network for multi-scale skin lesion image segmentation.Transformer 引导的自适网络用于多尺度皮肤病变图像分割。
Comput Biol Med. 2024 Feb;169:107846. doi: 10.1016/j.compbiomed.2023.107846. Epub 2023 Dec 23.
4
Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image.基于单目相机图像的道路车道和车道线检测中的交互式注意力学习
Sensors (Basel). 2023 Jul 20;23(14):6545. doi: 10.3390/s23146545.
5
DRTN: Dual Relation Transformer Network with feature erasure and contrastive learning for multi-label image classification.DRTN:用于多标签图像分类的具有特征擦除和对比学习的双关系Transformer网络。
Neural Netw. 2025 Jul;187:107309. doi: 10.1016/j.neunet.2025.107309. Epub 2025 Mar 3.
6
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.
7
The geometric attention-aware network for lane detection in complex road scenes.用于复杂道路场景中车道检测的几何注意感知网络。
PLoS One. 2021 Jul 15;16(7):e0254521. doi: 10.1371/journal.pone.0254521. eCollection 2021.
8
Aggregate global features into separable hierarchical lane detection transformer.将全局特征聚合到可分离的分层车道检测变换器中。
Sci Rep. 2025 Jan 22;15(1):2804. doi: 10.1038/s41598-025-86894-z.
9
Feature-Tuning Hierarchical Transformer via token communication and sample aggregation constraint for object re-identification.通过令牌通信和样本聚合约束对分层变压器进行特征微调以实现目标重识别
Neural Netw. 2025 Jul;187:107394. doi: 10.1016/j.neunet.2025.107394. Epub 2025 Mar 20.
10
Multi-Granularity Part Sampling Attention for Fine-Grained Visual Classification.用于细粒度视觉分类的多粒度部分采样注意力机制
IEEE Trans Image Process. 2024;33:4529-4542. doi: 10.1109/TIP.2024.3441813. Epub 2024 Aug 23.