• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

LST-BEV:为多视图3D目标检测生成长期时空鸟瞰图特征

LST-BEV: Generating a Long-Term Spatial-Temporal Bird's-Eye-View Feature for Multi-View 3D Object Detection.

作者信息

Feng Qijun, Zhao Chunyang, Liu Pengfei, Zhang Zhichao, Jin Yue, Tian Wanglin

机构信息

School of Information Science and Engineering, Shenyang Ligong University, Shenyang 110159, China.

Shenyang Institute of Automation Chinese Academy of Sciences, Shenyang 110169, China.

出版信息

Sensors (Basel). 2025 Jun 28;25(13):4040. doi: 10.3390/s25134040.

DOI:10.3390/s25134040
PMID:40648296
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12252236/
Abstract

This paper presents a novel multi-view 3D object detection framework, Long-Term Spatial-Temporal Bird's-Eye View (LST-BEV), designed to improve performance in autonomous driving. Traditional 3D detection relies on sensors like LiDAR, but visual perception using multi-camera systems is emerging as a more cost-effective solution. Existing methods struggle with capturing long-range dependencies and cross-task information due to limitations in attention mechanisms. To address this, we propose a Long-Range Cross-Task Detection Head (LRCH) to capture these dependencies and integrate cross-task information for accurate predictions. Additionally, we introduce the Long-Term Temporal Perception Module (LTPM), which efficiently extracts temporal features by combining Mamba and linear attention, overcoming challenges in temporal frame extraction. Experimental results in the nuScenes dataset demonstrate that our proposed LST-BEV outperforms its baseline (SA-BEVPool) by 2.1% mAP and 2.7% NDS, indicating a significant performance improvement.

摘要

本文提出了一种新颖的多视图3D目标检测框架,即长期时空鸟瞰视图(LST-BEV),旨在提高自动驾驶中的性能。传统的3D检测依赖于激光雷达等传感器,但使用多摄像头系统的视觉感知正成为一种更具成本效益的解决方案。由于注意力机制的限制,现有方法在捕获长距离依赖关系和跨任务信息方面存在困难。为了解决这个问题,我们提出了一种长距离跨任务检测头(LRCH)来捕获这些依赖关系,并集成跨任务信息以进行准确预测。此外,我们引入了长期时间感知模块(LTPM),它通过结合曼巴和线性注意力有效地提取时间特征,克服了时间帧提取中的挑战。在nuScenes数据集中的实验结果表明,我们提出的LST-BEV比其基线(SA-BEVPool)的平均精度均值(mAP)高2.1%,归一化检测分数(NDS)高2.7%,表明性能有显著提升。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/5c8d460bc31a/sensors-25-04040-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/c6d1dca963ce/sensors-25-04040-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/1020256e8de5/sensors-25-04040-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/2ea061d237c5/sensors-25-04040-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/73e8ec2c1c05/sensors-25-04040-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/5c8d460bc31a/sensors-25-04040-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/c6d1dca963ce/sensors-25-04040-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/1020256e8de5/sensors-25-04040-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/2ea061d237c5/sensors-25-04040-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/73e8ec2c1c05/sensors-25-04040-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b729/12252236/5c8d460bc31a/sensors-25-04040-g005.jpg

相似文献

1
LST-BEV: Generating a Long-Term Spatial-Temporal Bird's-Eye-View Feature for Multi-View 3D Object Detection.LST-BEV:为多视图3D目标检测生成长期时空鸟瞰图特征
Sensors (Basel). 2025 Jun 28;25(13):4040. doi: 10.3390/s25134040.
2
PolarFusion: A multi-modal fusion algorithm for 3D object detection based on polar coordinates.PolarFusion:一种基于极坐标的用于3D目标检测的多模态融合算法。
Neural Netw. 2025 Oct;190:107704. doi: 10.1016/j.neunet.2025.107704. Epub 2025 Jun 18.
3
Short-Term Memory Impairment短期记忆障碍
4
Spatio-temporal transformer and graph convolutional networks based traffic flow prediction.基于时空变换器和图卷积网络的交通流预测
Sci Rep. 2025 Jul 7;15(1):24299. doi: 10.1038/s41598-025-10287-5.
5
BEVFormer: Learning Bird's-Eye-View Representation From LiDAR-Camera Via Spatiotemporal Transformers.BEVFormer:通过时空变换器从激光雷达-相机学习鸟瞰视角表示
IEEE Trans Pattern Anal Mach Intell. 2024 Dec 10;PP. doi: 10.1109/TPAMI.2024.3515454.
6
DASNet a dual branch multi level attention sheep counting network.DASNet是一种双分支多级注意力羊只计数网络。
Sci Rep. 2025 Jul 2;15(1):23228. doi: 10.1038/s41598-025-97929-w.
7
Treatment of newly diagnosed glioblastoma in the elderly: a network meta-analysis.老年新诊断胶质母细胞瘤的治疗:一项网状Meta分析
Cochrane Database Syst Rev. 2020 Mar 23;3(3):CD013261. doi: 10.1002/14651858.CD013261.pub2.
8
Fast-BEV: A Fast and Strong Bird's-Eye View Perception Baseline.Fast-BEV:一种快速且强大的鸟瞰视角感知基线。
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):8665-8679. doi: 10.1109/TPAMI.2024.3414835. Epub 2024 Nov 6.
9
Treatment options for progression or recurrence of glioblastoma: a network meta-analysis.治疗胶质母细胞瘤进展或复发的选择:网络荟萃分析。
Cochrane Database Syst Rev. 2021 May 4;5(1):CD013579. doi: 10.1002/14651858.CD013579.pub2.
10
VMDU-net: a dual encoder multi-scale fusion network for polyp segmentation with Vision Mamba and Cross-Shape Transformer integration.VMDU-net:一种用于息肉分割的双编码器多尺度融合网络,集成了视觉曼巴和十字形变换器
Front Artif Intell. 2025 Jun 18;8:1557508. doi: 10.3389/frai.2025.1557508. eCollection 2025.

本文引用的文献

1
BEVFormer: Learning Bird's-Eye-View Representation From LiDAR-Camera Via Spatiotemporal Transformers.BEVFormer:通过时空变换器从激光雷达-相机学习鸟瞰视角表示
IEEE Trans Pattern Anal Mach Intell. 2024 Dec 10;PP. doi: 10.1109/TPAMI.2024.3515454.
2
A Survey of Visual Transformers.视觉Transformer综述
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7478-7498. doi: 10.1109/TNNLS.2022.3227717. Epub 2024 Jun 3.
3
Image Super-Resolution Using Deep Convolutional Networks.基于深度卷积网络的图像超分辨率重建。
IEEE Trans Pattern Anal Mach Intell. 2016 Feb;38(2):295-307. doi: 10.1109/TPAMI.2015.2439281.