• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

NDNet:通过邻域解耦进行空间多尺度表示学习以实现实时驾驶场景解析

NDNet: Spacewise Multiscale Representation Learning via Neighbor Decoupling for Real-Time Driving Scene Parsing.

作者信息

Li Shu, Yan Qingqing, Zhou Xun, Wang Deming, Liu Chengju, Chen Qijun

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7884-7898. doi: 10.1109/TNNLS.2022.3221745. Epub 2024 Jun 3.

DOI:10.1109/TNNLS.2022.3221745
PMID:36409808
Abstract

As a safety-critical application, autonomous driving requires high-quality semantic segmentation and real-time performance for deployment. Existing method commonly suffers from information loss and massive computational burden due to high-resolution input-output and multiscale learning scheme, which runs counter to the real-time requirements. In contrast to channelwise information modeling commonly adopted by modern networks, in this article, we propose a novel real-time driving scene parsing framework named NDNet from a novel perspective of spacewise neighbor decoupling (ND) and neighbor coupling (NC). We first define and implement the reversible operations called ND and NC, which realize lossless resolution conversion for complementary thumbnails sampling and collation to facilitate spatial modeling. Based on ND and NC, we further propose three modules, namely, local capturer and global dependence builder (LCGB), spacewise multiscale feature extractor (SMFE), and high-resolution semantic generator (HSG), which form the whole pipeline of NDNet. The LCGB serves as a stem block to preprocess the large-scale input for fast but lossless resolution reduction and extract initial features with global context. Then the SMFE is used for dense feature extraction and can obtain rich multiscale features in spatial dimension with less computational overhead. As for high-resolution semantic output, the HSG is designed for fast resolution reconstruction and adaptive semantic confusion amending. Experiments show the superiority of the proposed method. NDNet achieves the state-of-the-art performance on the Cityscapes dataset which reports 76.47% mIoU at 240+ frames/s and 78.8% mIoU at 150+ frames/s on the benchmark. Codes are available at https://github.com/LiShuTJ/NDNet.

摘要

作为一种安全关键型应用,自动驾驶在部署时需要高质量的语义分割和实时性能。由于高分辨率的输入输出和多尺度学习方案,现有方法通常存在信息丢失和巨大的计算负担,这与实时要求背道而驰。与现代网络普遍采用的通道级信息建模不同,在本文中,我们从空间邻域解耦(ND)和邻域耦合(NC)的新视角提出了一种名为NDNet的新型实时驾驶场景解析框架。我们首先定义并实现了名为ND和NC的可逆操作,它们实现了互补缩略图采样和整理的无损分辨率转换,以促进空间建模。基于ND和NC,我们进一步提出了三个模块,即局部捕获器和全局依赖性构建器(LCGB)、空间多尺度特征提取器(SMFE)和高分辨率语义生成器(HSG),它们构成了NDNet的整个流程。LCGB作为一个主干模块,对大规模输入进行预处理,以实现快速但无损的分辨率降低,并提取具有全局上下文的初始特征。然后,SMFE用于密集特征提取,并且可以在空间维度上以较少的计算开销获得丰富的多尺度特征。至于高分辨率语义输出,HSG旨在快速分辨率重建和自适应语义混淆修正。实验表明了所提方法的优越性。NDNet在Cityscapes数据集上达到了当前最优性能,在基准测试中,在240+帧/秒时报告的平均交并比为76.47%,在150+帧/秒时为78.8%。代码可在https://github.com/LiShuTJ/NDNet获取。

相似文献

1
NDNet: Spacewise Multiscale Representation Learning via Neighbor Decoupling for Real-Time Driving Scene Parsing.NDNet:通过邻域解耦进行空间多尺度表示学习以实现实时驾驶场景解析
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7884-7898. doi: 10.1109/TNNLS.2022.3221745. Epub 2024 Jun 3.
2
Small Object Augmentation of Urban Scenes for Real-Time Semantic Segmentation.用于实时语义分割的城市场景小目标增强
IEEE Trans Image Process. 2020 Mar 18. doi: 10.1109/TIP.2020.2976856.
3
LMFFNet: A Well-Balanced Lightweight Network for Fast and Accurate Semantic Segmentation.LMFFNet:用于快速准确语义分割的均衡轻量化网络。
IEEE Trans Neural Netw Learn Syst. 2023 Jun;34(6):3205-3219. doi: 10.1109/TNNLS.2022.3176493. Epub 2023 Jun 1.
4
Context and Spatial Feature Calibration for Real-Time Semantic Segmentation.用于实时语义分割的上下文和空间特征校准
IEEE Trans Image Process. 2023;32:5465-5477. doi: 10.1109/TIP.2023.3318967. Epub 2023 Oct 25.
5
Semantic segmentation of autonomous driving scenes based on multi-scale adaptive attention mechanism.基于多尺度自适应注意力机制的自动驾驶场景语义分割
Front Neurosci. 2023 Oct 19;17:1291674. doi: 10.3389/fnins.2023.1291674. eCollection 2023.
6
Real-Time Semantic Segmentation via a Densely Aggregated Bilateral Network.通过密集聚合双边网络实现实时语义分割
IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):381-392. doi: 10.1109/TNNLS.2023.3326665. Epub 2025 Jan 7.
7
MAFFNet: real-time multi-level attention feature fusion network with RGB-D semantic segmentation for autonomous driving.MAFFNet:用于自动驾驶的具有RGB-D语义分割的实时多级注意力特征融合网络
Appl Opt. 2022 Mar 20;61(9):2219-2229. doi: 10.1364/AO.449589.
8
LayerNet: A One-Step Layered Network for Semantic Segmentation at Night.LayerNet:一种用于夜间语义分割的单步分层网络。
IEEE Comput Graph Appl. 2023 Nov-Dec;43(6):9-21. doi: 10.1109/MCG.2023.3253167. Epub 2023 Nov 6.
9
CCNet: Criss-Cross Attention for Semantic Segmentation.CCNet:用于语义分割的交叉注意力。
IEEE Trans Pattern Anal Mach Intell. 2023 Jun;45(6):6896-6908. doi: 10.1109/TPAMI.2020.3007032. Epub 2023 May 5.
10
EPMF: Efficient Perception-Aware Multi-Sensor Fusion for 3D Semantic Segmentation.EPMF:用于3D语义分割的高效感知感知多传感器融合
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):8258-8273. doi: 10.1109/TPAMI.2024.3402232. Epub 2024 Nov 6.