• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于序分类的序列多尺度时空特征提取与深度估计。

Multi-Scale Spatio-Temporal Feature Extraction and Depth Estimation from Sequences by Ordinal Classification.

机构信息

School of Digital Media & Design Arts, Beijing University of Posts and Telecommunications, Beijing 100876, China.

Beijing Key Laboratory of Network System and Network Culture, Beijing, 100876, China.

出版信息

Sensors (Basel). 2020 Apr 1;20(7):1979. doi: 10.3390/s20071979.

DOI:10.3390/s20071979
PMID:32244820
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7180695/
Abstract

Depth estimation is a key problem in 3D computer vision and has a wide variety of applications. In this paper we explore whether deep learning network can predict depth map accurately by learning multi-scale spatio-temporal features from sequences and recasting the depth estimation from a regression task to an ordinal classification task. We design an encoder-decoder network with several multi-scale strategies to improve its performance and extract spatio-temporal features with ConvLSTM. The results of our experiments show that the proposed method has an improvement of almost 10% in error metrics and up to 2% in accuracy metrics. The results also tell us that extracting spatio-temporal features can dramatically improve the performance in depth estimation task. We consider to extend this work to a self-supervised manner to get rid of the dependence on large-scale labeled data.

摘要

深度估计是三维计算机视觉中的一个关键问题,具有广泛的应用。在本文中,我们探讨了深度学习网络是否可以通过从序列中学习多尺度时空特征并将深度估计从回归任务重新定义为有序分类任务,从而准确地预测深度图。我们设计了一个具有多种多尺度策略的编码器-解码器网络,以提高其性能并使用 ConvLSTM 提取时空特征。实验结果表明,所提出的方法在误差指标上提高了近 10%,在精度指标上提高了 2%。结果还表明,提取时空特征可以显著提高深度估计任务的性能。我们考虑将这项工作扩展到自监督的方式,以摆脱对大规模标记数据的依赖。

相似文献

1
Multi-Scale Spatio-Temporal Feature Extraction and Depth Estimation from Sequences by Ordinal Classification.基于序分类的序列多尺度时空特征提取与深度估计。
Sensors (Basel). 2020 Apr 1;20(7):1979. doi: 10.3390/s20071979.
2
Automatic Extraction of Water and Shadow from SAR Images Based on a Multi-Resolution Dense Encoder and Decoder Network.基于多分辨率密集编码器和解码器网络的合成孔径雷达图像水和阴影自动提取
Sensors (Basel). 2019 Aug 16;19(16):3576. doi: 10.3390/s19163576.
3
Deep learning with 4D spatio-temporal data representations for OCT-based force estimation.基于 OCT 的力估计的 4D 时空数据表示的深度学习。
Med Image Anal. 2020 Aug;64:101730. doi: 10.1016/j.media.2020.101730. Epub 2020 May 23.
4
Spatio-Temporal Convolutional LSTMs for Tumor Growth Prediction by Learning 4D Longitudinal Patient Data.基于 4D 纵向患者数据学习的时空卷积长短期记忆模型在肿瘤生长预测中的应用。
IEEE Trans Med Imaging. 2020 Apr;39(4):1114-1126. doi: 10.1109/TMI.2019.2943841. Epub 2019 Sep 25.
5
Fast Depth Estimation in a Single Image Using Lightweight Efficient Neural Network.基于轻量级高效神经网络的单图像快速深度估计。
Sensors (Basel). 2019 Oct 13;19(20):4434. doi: 10.3390/s19204434.
6
Spatio-temporal layers based intra-operative stereo depth estimation network via hierarchical prediction and progressive training.基于时空层的术中立体深度估计网络,通过分层预测和渐进式训练。
Comput Methods Programs Biomed. 2024 Feb;244:107937. doi: 10.1016/j.cmpb.2023.107937. Epub 2023 Nov 22.
7
Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance.基于迁移学习和表面法向导引的卓越单目深度估计。
Sensors (Basel). 2020 Aug 27;20(17):4856. doi: 10.3390/s20174856.
8
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.
9
Audio-visual multi-modality driven hybrid feature learning model for crowd analysis and classification.用于人群分析与分类的视听多模态驱动混合特征学习模型
Math Biosci Eng. 2023 May 25;20(7):12529-12561. doi: 10.3934/mbe.2023558.
10
Skeleton-Based Spatio-Temporal U-Network for 3D Human Pose Estimation in Video.基于骨架的时空 U-Net 网络用于视频中的 3D 人体姿态估计。
Sensors (Basel). 2022 Mar 28;22(7):2573. doi: 10.3390/s22072573.

本文引用的文献

1
Incremental Pose Map Optimization for Monocular Vision SLAM Based on Similarity Transformation.基于相似变换的单目视觉 SLAM 增量位姿图优化。
Sensors (Basel). 2019 Nov 13;19(22):4945. doi: 10.3390/s19224945.
2
Fast Depth Estimation in a Single Image Using Lightweight Efficient Neural Network.基于轻量级高效神经网络的单图像快速深度估计。
Sensors (Basel). 2019 Oct 13;19(20):4434. doi: 10.3390/s19204434.
3
SemanticDepth: Fusing Semantic Segmentation and Monocular Depth Estimation for Enabling Autonomous Driving in Roads without Lane Lines.
语义深度:融合语义分割与单目深度估计以实现无车道线道路上的自动驾驶
Sensors (Basel). 2019 Jul 22;19(14):3224. doi: 10.3390/s19143224.
4
Monocular Robust Depth Estimation Vision System for Robotic Tasks Interventions in Metallic Targets.用于金属目标机器人任务干预的单目稳健深度估计视觉系统。
Sensors (Basel). 2019 Jul 22;19(14):3220. doi: 10.3390/s19143220.
5
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.
6
Unsupervised Learning for Depth, Ego-Motion, and Optical Flow Estimation Using Coupled Consistency Conditions.使用耦合一致性条件的深度、自我运动和光流估计的无监督学习
Sensors (Basel). 2019 May 29;19(11):2459. doi: 10.3390/s19112459.
7
Depth Estimation and Semantic Segmentation from a Single RGB Image Using a Hybrid Convolutional Neural Network.使用混合卷积神经网络从单张RGB图像进行深度估计和语义分割
Sensors (Basel). 2019 Apr 15;19(8):1795. doi: 10.3390/s19081795.
8
Predicting Depth from Single RGB Images with Pyramidal Three-Streamed Networks.基于金字塔三流式网络的单 RGB 图像深度预测。
Sensors (Basel). 2019 Feb 6;19(3):667. doi: 10.3390/s19030667.
9
Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras.运动算法和智能相机硬件架构的深度
Sensors (Basel). 2018 Dec 23;19(1):53. doi: 10.3390/s19010053.
10
SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.SegNet:一种用于图像分割的深度卷积编解码器架构。
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495. doi: 10.1109/TPAMI.2016.2644615. Epub 2017 Jan 2.