Department of Computer Science, Technical University of München, Boltzmanstrasse 3, Munich, Germany.
IEEE Trans Pattern Anal Mach Intell. 2012 Mar;34(3):493-505. doi: 10.1109/TPAMI.2011.150.
We propose a probabilistic formulation of joint silhouette extraction and 3D reconstruction given a series of calibrated 2D images. Instead of segmenting each image separately in order to construct a 3D surface consistent with the estimated silhouettes, we compute the most probable 3D shape that gives rise to the observed color information. The probabilistic framework, based on Bayesian inference, enables robust 3D reconstruction by optimally taking into account the contribution of all views. We solve the arising maximum a posteriori shape inference in a globally optimal manner by convex relaxation techniques in a spatially continuous representation. For an interactively provided user input in the form of scribbles specifying foreground and background regions, we build corresponding color distributions as multivariate Gaussians and find a volume occupancy that best fits to this data in a variational sense. Compared to classical methods for silhouette-based multiview reconstruction, the proposed approach does not depend on initialization and enjoys significant resilience to violations of the model assumptions due to background clutter, specular reflections, and camera sensor perturbations. In experiments on several real-world data sets, we show that exploiting a silhouette coherency criterion in a multiview setting allows for dramatic improvements of silhouette quality over independent 2D segmentations without any significant increase of computational efforts. This results in more accurate visual hull estimation, needed by a multitude of image-based modeling approaches. We made use of recent advances in parallel computing with a GPU implementation of the proposed method generating reconstructions on volume grids of more than 20 million voxels in up to 4.41 seconds.
我们提出了一种联合轮廓提取和 3D 重建的概率公式,给定一系列校准的 2D 图像。我们不是为了构建与估计轮廓一致的 3D 表面而分别对每个图像进行分割,而是计算最有可能的 3D 形状,从而产生观察到的颜色信息。基于贝叶斯推理的概率框架通过最优地考虑所有视图的贡献,实现了稳健的 3D 重建。我们通过在空间连续表示中使用凸松弛技术,以全局最优的方式解决了出现的最大后验形状推断问题。对于以指定前景和背景区域的手写输入的形式提供的交互输入,我们构建相应的颜色分布作为多元高斯,并找到最佳的体积占有率,以便在变分意义上最好地适应此数据。与基于轮廓的多视图重建的经典方法相比,所提出的方法不依赖于初始化,并且由于背景杂波、镜面反射和相机传感器干扰等原因对违反模型假设具有显著的弹性。在对几个真实世界数据集的实验中,我们表明,在多视图设置中利用轮廓一致性准则可以在不显著增加计算工作量的情况下,显著提高轮廓质量,而无需任何显著增加计算工作量。这导致更准确的基于图像的建模方法所需的视觉外壳估计。我们利用了 GPU 实现的并行计算的最新进展,该方法在高达 4.41 秒的时间内生成超过 2000 万个体素的体积网格的重建。