• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过 2D/3D 标签传递实现户外数据集的一致语义标注。

Consistent Semantic Annotation of Outdoor Datasets via 2D/3D Label Transfer.

机构信息

School of Informatics, University of Edinburgh, Edinburgh EH8 9AB, UK.

出版信息

Sensors (Basel). 2018 Jul 12;18(7):2249. doi: 10.3390/s18072249.

DOI:10.3390/s18072249
PMID:30002334
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6069386/
Abstract

The advance of scene understanding methods based on machine learning relies on the availability of large ground truth datasets, which are essential for their training and evaluation. Construction of such datasets with imagery from real sensor data however typically requires much manual annotation of semantic regions in the data, delivered by substantial human labour. To speed up this process, we propose a framework for semantic annotation of scenes captured by moving camera(s), e.g., mounted on a vehicle or robot. It makes use of an available 3D model of the traversed scene to project segmented 3D objects into each camera frame to obtain an initial annotation of the associated 2D image, which is followed by manual refinement by the user. The refined annotation can be transferred to the next consecutive frame using optical flow estimation. We have evaluated the efficiency of the proposed framework during the production of a labelled outdoor dataset. The analysis of annotation times shows that up to 43% less effort is required on average, and the consistency of the labelling is also improved.

摘要

基于机器学习的场景理解方法的进展依赖于大规模的地面实况数据集的可用性,这些数据集对于它们的训练和评估至关重要。然而,使用真实传感器数据的图像构建这样的数据集通常需要对数据中的语义区域进行大量的手动标注,这需要大量的人力劳动。为了加快这个过程,我们提出了一个用于移动摄像机(例如安装在车辆或机器人上的摄像机)拍摄的场景的语义标注的框架。它利用可获得的场景的 3D 模型将分割的 3D 物体投影到每个摄像机帧中,以获得相关 2D 图像的初始标注,然后由用户进行手动细化。细化后的标注可以使用光流估计转移到下一个连续的帧。我们已经在一个标注的户外数据集的制作过程中评估了所提出的框架的效率。标注时间的分析表明,平均需要的工作量减少了 43%,标注的一致性也得到了提高。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/94bb5521b14c/sensors-18-02249-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/a0c245cd60b4/sensors-18-02249-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/da629dcb5fc0/sensors-18-02249-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/84379be1c1e5/sensors-18-02249-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/c2836dd6a331/sensors-18-02249-g0A4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/9f3b346b19a4/sensors-18-02249-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/d89728fdb1d8/sensors-18-02249-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/17674986eb36/sensors-18-02249-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/2fcd3ff720f4/sensors-18-02249-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/9a0d23055607/sensors-18-02249-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/b346da4cace0/sensors-18-02249-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/695f488bdf45/sensors-18-02249-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/e8a7735c7a35/sensors-18-02249-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/c2b47c21500c/sensors-18-02249-g009a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/94bb5521b14c/sensors-18-02249-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/a0c245cd60b4/sensors-18-02249-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/da629dcb5fc0/sensors-18-02249-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/84379be1c1e5/sensors-18-02249-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/c2836dd6a331/sensors-18-02249-g0A4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/9f3b346b19a4/sensors-18-02249-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/d89728fdb1d8/sensors-18-02249-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/17674986eb36/sensors-18-02249-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/2fcd3ff720f4/sensors-18-02249-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/9a0d23055607/sensors-18-02249-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/b346da4cace0/sensors-18-02249-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/695f488bdf45/sensors-18-02249-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/e8a7735c7a35/sensors-18-02249-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/c2b47c21500c/sensors-18-02249-g009a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1163/6069386/94bb5521b14c/sensors-18-02249-g010.jpg

相似文献

1
Consistent Semantic Annotation of Outdoor Datasets via 2D/3D Label Transfer.通过 2D/3D 标签传递实现户外数据集的一致语义标注。
Sensors (Basel). 2018 Jul 12;18(7):2249. doi: 10.3390/s18072249.
2
Efficient 3D Scene Semantic Segmentation via Active Learning on Rendered 2D Images.基于渲染二维图像的主动学习的高效三维场景语义分割。
IEEE Trans Image Process. 2023;32:3521-3535. doi: 10.1109/TIP.2023.3286708. Epub 2023 Jun 29.
3
Transfer Learning Based Semantic Segmentation for 3D Object Detection from Point Cloud.基于迁移学习的点云三维目标检测语义分割。
Sensors (Basel). 2021 Jun 8;21(12):3964. doi: 10.3390/s21123964.
4
GTCreator: a flexible annotation tool for image-based datasets.GTCreator:一个用于基于图像数据集的灵活标注工具。
Int J Comput Assist Radiol Surg. 2019 Feb;14(2):191-201. doi: 10.1007/s11548-018-1864-x. Epub 2018 Sep 25.
5
MAIA-A machine learning assisted image annotation method for environmental monitoring and exploration.MAIA-一种用于环境监测和探索的机器学习辅助图像标注方法。
PLoS One. 2018 Nov 16;13(11):e0207498. doi: 10.1371/journal.pone.0207498. eCollection 2018.
6
Deep learning of the sectional appearances of 3D CT images for anatomical structure segmentation based on an FCN voting method.基于 FCN 投票方法的三维 CT 图像节段外观的深度学习用于解剖结构分割。
Med Phys. 2017 Oct;44(10):5221-5233. doi: 10.1002/mp.12480. Epub 2017 Aug 31.
7
ODMedit: uniform semantic annotation for data integration in medicine based on a public metadata repository.ODMedit:基于公共元数据存储库的医学数据集成的统一语义标注。
BMC Med Res Methodol. 2016 Jun 1;16:65. doi: 10.1186/s12874-016-0164-9.
8
A Robust 3D-2D Interactive Tool for Scene Segmentation and Annotation.一种用于场景分割与标注的强大3D-2D交互式工具。
IEEE Trans Vis Comput Graph. 2018 Dec;24(12):3005-3018. doi: 10.1109/TVCG.2017.2772238. Epub 2017 Nov 20.
9
Bias-Eliminated Semantic Refinement for Any-Shot Learning.无偏语义细化的任意-shot 学习。
IEEE Trans Image Process. 2022;31:2229-2244. doi: 10.1109/TIP.2022.3152631. Epub 2022 Mar 8.
10
An annotation and modeling schema for prescription regimens.处方用药方案的注释与建模架构
J Biomed Semantics. 2019 May 31;10(1):10. doi: 10.1186/s13326-019-0201-9.

引用本文的文献

1
Building Human Visual Attention Map for Construction Equipment Teleoperation.构建用于建筑设备远程操作的人类视觉注意力地图。
Front Neurosci. 2022 Jun 10;16:895126. doi: 10.3389/fnins.2022.895126. eCollection 2022.

本文引用的文献

1
A Robust 3D-2D Interactive Tool for Scene Segmentation and Annotation.一种用于场景分割与标注的强大3D-2D交互式工具。
IEEE Trans Vis Comput Graph. 2018 Dec;24(12):3005-3018. doi: 10.1109/TVCG.2017.2772238. Epub 2017 Nov 20.
2
SLIC superpixels compared to state-of-the-art superpixel methods.SLIC 超像素与最先进的超像素方法比较。
IEEE Trans Pattern Anal Mach Intell. 2012 Nov;34(11):2274-82. doi: 10.1109/TPAMI.2012.120.
3
Nonparametric Scene Parsing via Label Transfer.基于标签转移的非参数场景解析。
IEEE Trans Pattern Anal Mach Intell. 2011 Dec;33(12):2368-82. doi: 10.1109/TPAMI.2011.131. Epub 2011 Jun 30.