IEEE Trans Pattern Anal Mach Intell. 2011 Oct;33(10):1952-61. doi: 10.1109/TPAMI.2011.123. Epub 2011 Jun 16.
Classical face recognition techniques have been successful at operating under well-controlled conditions; however, they have difficulty in robustly performing recognition in uncontrolled real-world scenarios where variations in pose, illumination, and expression are encountered. In this paper, we propose a new method for real-world unconstrained pose-invariant face recognition. We first construct a 3D model for each subject in our database using only a single 2D image by applying the 3D Generic Elastic Model (3D GEM) approach. These 3D models comprise an intermediate gallery database from which novel 2D pose views are synthesized for matching. Before matching, an initial estimate of the pose of the test query is obtained using a linear regression approach based on automatic facial landmark annotation. Each 3D model is subsequently rendered at different poses within a limited search space about the estimated pose, and the resulting images are matched against the test query. Finally, we compute the distances between the synthesized images and test query by using a simple normalized correlation matcher to show the effectiveness of our pose synthesis method to real-world data. We present convincing results on challenging data sets and video sequences demonstrating high recognition accuracy under controlled as well as unseen, uncontrolled real-world scenarios using a fast implementation.
经典的人脸识别技术在控制良好的条件下运行得非常成功;然而,在遇到姿势、光照和表情变化的不受控制的真实场景中,它们很难稳健地进行识别。在本文中,我们提出了一种新的用于真实世界无约束姿态不变人脸识别的方法。我们首先通过应用 3D 通用弹性模型(3D GEM)方法,仅使用单个 2D 图像为数据库中的每个主体构建 3D 模型。这些 3D 模型构成了一个中间库,从中可以合成新的 2D 姿态视图进行匹配。在匹配之前,使用基于自动面部地标注释的线性回归方法获得测试查询的初始姿势估计。随后,在估计姿势的有限搜索空间内,将每个 3D 模型渲染到不同的姿势,然后将生成的图像与测试查询进行匹配。最后,我们通过使用简单的归一化相关匹配器计算合成图像与测试查询之间的距离,以展示我们的姿态合成方法对真实世界数据的有效性。我们在具有挑战性的数据集和视频序列上呈现了令人信服的结果,展示了在使用快速实现的受控和未见过的不受控制的真实场景下的高识别准确性。