School of Computer Science and Engineering, South China University of Technology, Guangzhou, China.
Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong.
IEEE Trans Image Process. 2018;27(1):164-178. doi: 10.1109/TIP.2017.2736603.
Stitching videos captured by hand-held mobile cameras can essentially enhance entertainment experience of ordinary users. However, such videos usually contain heavy shakiness and large parallax, which are challenging to stitch. In this paper, we propose a novel approach of video stitching and stabilization for videos captured by mobile devices. The main component of our method is a unified video stitching and stabilization optimization that computes stitching and stabilization simultaneously rather than does each one individually. In this way, we can obtain the best stitching and stabilization results relative to each other without any bias to one of them. To make the optimization robust, we propose a method to identify background of input videos, and also common background of them. This allows us to apply our optimization on background regions only, which is the key to handle large parallax problem. Since stitching relies on feature matches between input videos, and there inevitably exist false matches, we thus propose a method to distinguish between right and false matches, and encapsulate the false match elimination scheme and our optimization into a loop, to prevent the optimization from being affected by bad feature matches. We test the proposed approach on videos that are causally captured by smartphones when walking along busy streets, and use stitching and stability scores to evaluate the produced panoramic videos quantitatively. Experiments on a diverse of examples show that our results are much better than (challenging cases) or at least on par with (simple cases) the results of previous approaches.Stitching videos captured by hand-held mobile cameras can essentially enhance entertainment experience of ordinary users. However, such videos usually contain heavy shakiness and large parallax, which are challenging to stitch. In this paper, we propose a novel approach of video stitching and stabilization for videos captured by mobile devices. The main component of our method is a unified video stitching and stabilization optimization that computes stitching and stabilization simultaneously rather than does each one individually. In this way, we can obtain the best stitching and stabilization results relative to each other without any bias to one of them. To make the optimization robust, we propose a method to identify background of input videos, and also common background of them. This allows us to apply our optimization on background regions only, which is the key to handle large parallax problem. Since stitching relies on feature matches between input videos, and there inevitably exist false matches, we thus propose a method to distinguish between right and false matches, and encapsulate the false match elimination scheme and our optimization into a loop, to prevent the optimization from being affected by bad feature matches. We test the proposed approach on videos that are causally captured by smartphones when walking along busy streets, and use stitching and stability scores to evaluate the produced panoramic videos quantitatively. Experiments on a diverse of examples show that our results are much better than (challenging cases) or at least on par with (simple cases) the results of previous approaches.
手持移动设备拍摄的视频拼接可以极大地提升普通用户的娱乐体验。然而,这些视频通常存在严重的抖动和大视差,拼接起来具有挑战性。在本文中,我们提出了一种用于移动设备拍摄视频的拼接和稳定化的新方法。我们方法的主要组成部分是一个统一的视频拼接和稳定化优化,它可以同时计算拼接和稳定化,而不是分别进行。这样,我们可以在彼此之间获得最佳的拼接和稳定化结果,而不会对其中任何一个产生偏见。为了使优化具有鲁棒性,我们提出了一种方法来识别输入视频的背景,以及它们的常见背景。这使我们能够仅在背景区域应用我们的优化,这是处理大视差问题的关键。由于拼接依赖于输入视频之间的特征匹配,并且不可避免地存在错误匹配,因此我们提出了一种区分正确和错误匹配的方法,并将错误匹配消除方案和我们的优化封装在一个循环中,以防止优化受到不良特征匹配的影响。我们在智能手机沿着繁忙街道行走时拍摄的视频上测试了所提出的方法,并使用拼接和稳定性得分对生成的全景视频进行定量评估。在各种示例上的实验表明,我们的结果明显优于(具有挑战性的情况)或至少与(简单的情况)以前的方法相当。