• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

KITTI-360:用于二维和三维城市场景理解的新型数据集和基准

KITTI-360: A Novel Dataset and Benchmarks for Urban Scene Understanding in 2D and 3D.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3292-3310. doi: 10.1109/TPAMI.2022.3179507. Epub 2023 Feb 3.

DOI:10.1109/TPAMI.2022.3179507
PMID:35648872
Abstract

For the last few decades, several major subfields of artificial intelligence including computer vision, graphics, and robotics have progressed largely independently from each other. Recently, however, the community has realized that progress towards robust intelligent systems such as self-driving cars requires a concerted effort across the different fields. This motivated us to develop KITTI-360, successor of the popular KITTI dataset. KITTI-360 is a suburban driving dataset which comprises richer input modalities, comprehensive semantic instance annotations and accurate localization to facilitate research at the intersection of vision, graphics and robotics. For efficient annotation, we created a tool to label 3D scenes with bounding primitives and developed a model that transfers this information into the 2D image domain, resulting in over 150k images and 1B 3D points with coherent semantic instance annotations across 2D and 3D. Moreover, we established benchmarks and baselines for several tasks relevant to mobile perception, encompassing problems from computer vision, graphics, and robotics on the same dataset, e.g., semantic scene understanding, novel view synthesis and semantic SLAM. KITTI-360 will enable progress at the intersection of these research areas and thus contribute towards solving one of today's grand challenges: the development of fully autonomous self-driving systems.

摘要

在过去的几十年里,人工智能的几个主要子领域,包括计算机视觉、图形和机器人技术,在很大程度上是相互独立发展的。然而,最近,该领域的研究人员已经意识到,要实现自动驾驶汽车等强大智能系统的进步,需要在不同领域进行协同努力。这促使我们开发了 KITTI-360,这是广受欢迎的 KITTI 数据集的后续版本。KITTI-360 是一个郊区驾驶数据集,它包含更丰富的输入模式、全面的语义实例注释和精确的定位,以促进视觉、图形和机器人技术交叉领域的研究。为了高效地进行注释,我们创建了一个工具来用边界框标注 3D 场景,并开发了一个模型,将这些信息转换到 2D 图像域中,从而在 2D 和 3D 之间产生了超过 150k 张图像和 10 亿个具有连贯语义实例注释的 3D 点。此外,我们为几个与移动感知相关的任务建立了基准和基线,涵盖了来自计算机视觉、图形和机器人技术的问题,例如语义场景理解、新视图合成和语义 SLAM。KITTI-360 将促进这些研究领域的交叉发展,从而有助于解决当今的一个重大挑战:开发完全自主的自动驾驶系统。

相似文献

1
KITTI-360: A Novel Dataset and Benchmarks for Urban Scene Understanding in 2D and 3D.KITTI-360:用于二维和三维城市场景理解的新型数据集和基准
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3292-3310. doi: 10.1109/TPAMI.2022.3179507. Epub 2023 Feb 3.
2
The ApolloScape Open Dataset for Autonomous Driving and Its Application.阿波罗景观开放数据集在自动驾驶中的应用
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2702-2719. doi: 10.1109/TPAMI.2019.2926463. Epub 2019 Jul 2.
3
A Robust 3D-2D Interactive Tool for Scene Segmentation and Annotation.一种用于场景分割与标注的强大3D-2D交互式工具。
IEEE Trans Vis Comput Graph. 2018 Dec;24(12):3005-3018. doi: 10.1109/TVCG.2017.2772238. Epub 2017 Nov 20.
4
Visual Object Recognition with 3D-Aware Features in KITTI Urban Scenes.基于KITTI城市场景中3D感知特征的视觉目标识别
Sensors (Basel). 2015 Apr 20;15(4):9228-50. doi: 10.3390/s150409228.
5
Lowis3D: Language-Driven Open-World Instance-Level 3D Scene Understanding.Lowis3D:语言驱动的开放世界实例级3D场景理解
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):8517-8533. doi: 10.1109/TPAMI.2024.3410324. Epub 2024 Nov 6.
6
Road and Railway Smart Mobility: A High-Definition Ground Truth Hybrid Dataset.道路与铁路智能交通:高清地面实况混合数据集。
Sensors (Basel). 2022 May 22;22(10):3922. doi: 10.3390/s22103922.
7
DASGIL: Domain Adaptation for Semantic and Geometric-Aware Image-Based Localization.DASGIL:基于图像的语义和几何感知定位的域适应
IEEE Trans Image Process. 2021;30:1342-1353. doi: 10.1109/TIP.2020.3043875. Epub 2020 Dec 23.
8
3D Object Detection with SLS-Fusion Network in Foggy Weather Conditions.雾天条件下基于SLS融合网络的3D目标检测
Sensors (Basel). 2021 Oct 9;21(20):6711. doi: 10.3390/s21206711.
9
Representations and Benchmarking of Modern Visual SLAM Systems.现代视觉 SLAM 系统的表示与基准测试
Sensors (Basel). 2020 Apr 30;20(9):2572. doi: 10.3390/s20092572.
10
A Curriculum Domain Adaptation Approach to the Semantic Segmentation of Urban Scenes.一种用于城市场景语义分割的课程领域适应方法。
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):1823-1841. doi: 10.1109/TPAMI.2019.2903401. Epub 2019 Mar 6.

引用本文的文献

1
Metric scale non-fixed obstacles distance estimation using a 3D map and a monocular camera.使用三维地图和单目相机进行公制尺度非固定障碍物距离估计。
Front Robot AI. 2025 Jun 12;12:1560342. doi: 10.3389/frobt.2025.1560342. eCollection 2025.
2
SGF-SLAM: Semantic Gaussian Filtering SLAM for Urban Road Environments.SGF-SLAM:用于城市道路环境的语义高斯滤波同步定位与地图构建
Sensors (Basel). 2025 Jun 7;25(12):3602. doi: 10.3390/s25123602.
3
CMHT autonomous dataset: A multi-sensor dataset including radar and IR for autonomous driving.CMHT自主数据集:一个用于自动驾驶的包含雷达和红外的多传感器数据集。
Data Brief. 2025 Apr 11;60:111552. doi: 10.1016/j.dib.2025.111552. eCollection 2025 Jun.
4
Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection.辅助监督深度辅助变压器和跨模态注意力融合在单目3D目标检测中的应用。
PeerJ Comput Sci. 2025 Jan 28;11:e2656. doi: 10.7717/peerj-cs.2656. eCollection 2025.
5
A travelable area boundary dataset for visual navigation of field robots.用于田间机器人视觉导航的可通行区域边界数据集。
Sci Data. 2025 Jan 28;12(1):164. doi: 10.1038/s41597-025-04457-3.
6
Neural Radiance Fields for Fisheye Driving Scenes Using Edge-Aware Integrated Depth Supervision.使用边缘感知集成深度监督的鱼眼驾驶场景神经辐射场
Sensors (Basel). 2024 Oct 22;24(21):6790. doi: 10.3390/s24216790.
7
Large-Scale Indoor Visual-Geometric Multimodal Dataset and Benchmark for Novel View Synthesis.用于新视图合成的大规模室内视觉几何多模态数据集与基准
Sensors (Basel). 2024 Sep 6;24(17):5798. doi: 10.3390/s24175798.
8
STRIDE: Street View-based Environmental Feature Detection and Pedestrian Collision Prediction.STRIDE:基于街景的环境特征检测与行人碰撞预测
IEEE Int Conf Comput Vis Workshops. 2023 Oct;2023:3222-3234. doi: 10.1109/iccvw60793.2023.00347. Epub 2023 Dec 25.
9
Open-Vocabulary Predictive World Models from Sensor Observations.基于传感器观测的开放词汇预测世界模型
Sensors (Basel). 2024 Jul 21;24(14):4735. doi: 10.3390/s24144735.
10
DOT-SLAM: A Stereo Visual Simultaneous Localization and Mapping (SLAM) System with Dynamic Object Tracking Based on Graph Optimization.DOT-SLAM:一种基于图优化的具有动态目标跟踪功能的立体视觉同步定位与地图构建(SLAM)系统。
Sensors (Basel). 2024 Jul 18;24(14):4676. doi: 10.3390/s24144676.