Favaro P, Soatto S
IEEE Trans Pattern Anal Mach Intell. 2005 Mar;27(3):406-417. doi: 10.1109/TPAMI.2005.43.
We introduce a novel approach to shape from defocus, i.e., the problem of inferring the three-dimensional (3D) geometry of a scene from a collection of defocused images. Typically, in shape from defocus, the task of extracting geometry also requires deblurring the given images. A common approach to bypass this task relies on approximating the scene locally by a plane parallel to the image (the so-called equifocal assumption). We show that this approximation is indeed not necessary, as one can estimate 3D geometry while avoiding deblurring without strong assumptions on the scene. Solving the problem of shape from defocus requires modeling how light interacts with the optics before reaching the imaging surface. This interaction is described by the so-called point spread function (PSF). When the form of the PSF is known, we propose an optimal method to infer 3D geometry from defocused images that involves computing orthogonal operators which are regularized via functional singular value decomposition. When the form of the PSF is unknown, we propose a simple and efficient method that first learns a set of projection operators from blurred images and then uses these operators to estimate the 3D geometry of the scene from novel blurred images. Our experiments on both real and synthetic images show that the performance of the algorithm is relatively insensitive to the form of the PSF. Our general approach is to minimize the Euclidean norm of the difference between the estimated images and the observed images. The method is geometric in that we reduce the minimization to performing projections onto linear subspaces, by using inner product structures on both infinite and finite-dimensional Hilbert spaces. Both proposed algorithms involve only simple matrix-vector multiplications which can be implemented in real-time.
我们介绍了一种从散焦中恢复形状的新方法,即从一组散焦图像推断场景三维(3D)几何形状的问题。通常,在从散焦中恢复形状时,提取几何形状的任务还需要对给定图像进行去模糊处理。一种绕过此任务的常见方法依赖于用平行于图像的平面局部逼近场景(所谓的等焦假设)。我们表明这种逼近实际上并非必要,因为在不对场景做强假设的情况下,人们可以在避免去模糊的同时估计3D几何形状。解决从散焦中恢复形状的问题需要对光在到达成像表面之前与光学器件的相互作用进行建模。这种相互作用由所谓的点扩散函数(PSF)描述。当PSF的形式已知时,我们提出一种从散焦图像推断3D几何形状的最优方法,该方法涉及计算通过函数奇异值分解进行正则化的正交算子。当PSF的形式未知时,我们提出一种简单有效的方法,该方法首先从模糊图像中学习一组投影算子,然后使用这些算子从新的模糊图像中估计场景的3D几何形状。我们在真实图像和合成图像上的实验表明,该算法的性能对PSF的形式相对不敏感。我们的一般方法是最小化估计图像与观测图像之间差异的欧几里得范数。该方法是几何方法,因为我们通过在无限维和有限维希尔伯特空间上使用内积结构,将最小化问题简化为在线性子空间上进行投影。所提出的两种算法都只涉及简单的矩阵 - 向量乘法,可实时实现。