Zhang Guofeng, Dong Zilong, Jia Jiaya, Wan Liang, Wong Tien-Tsin, Bao Hujun
State Key Lab of CA&GC, Zijingang Campus, Zhejiang University, Hangzhou 310058, P.R. China.
IEEE Trans Vis Comput Graph. 2009 Sep-Oct;15(5):828-40. doi: 10.1109/TVCG.2009.47.
Compared to still image editing, content-based video editing faces the additional challenges of maintaining the spatiotemporal consistency with respect to geometry. This brings up difficulties of seamlessly modifying video content, for instance, inserting or removing an object. In this paper, we present a new video editing system for creating spatiotemporally consistent and visually appealing refilming effects. Unlike the typical filming practice, our system requires no labor-intensive construction of 3D models/surfaces mimicking the real scene. Instead, it is based on an unsupervised inference of view-dependent depth maps for all video frames. We provide interactive tools requiring only a small amount of user input to perform elementary video content editing, such as separating video layers, completing background scene, and extracting moving objects. These tools can be utilized to produce a variety of visual effects in our system, including but not limited to video composition, "predator" effect, bullet-time, depth-of-field, and fog synthesis. Some of the effects can be achieved in real time.
与静态图像编辑相比,基于内容的视频编辑在保持几何形状的时空一致性方面面临额外挑战。这给无缝修改视频内容带来了困难,例如插入或移除一个物体。在本文中,我们提出了一种新的视频编辑系统,用于创建时空一致且视觉上吸引人的重新拍摄效果。与典型的拍摄实践不同,我们的系统不需要对模仿真实场景的3D模型/表面进行费力的构建。相反,它基于对所有视频帧的视图相关深度图的无监督推断。我们提供了仅需少量用户输入就能执行基本视频内容编辑的交互式工具,例如分离视频层、完成背景场景以及提取移动物体。这些工具可用于在我们的系统中产生各种视觉效果,包括但不限于视频合成、“掠食者”效果、子弹时间、景深和雾合成。其中一些效果可以实时实现。