Verma Monu, Vipparthi Santosh Kumar, Singh Girdhari, Murala Subrahmanyam
IEEE Trans Image Process. 2019 Sep 19. doi: 10.1109/TIP.2019.2912358.
Unlike prevalent facial expressions, micro expressions have subtle, involuntary muscle movements which are short-lived in nature. These minute muscle movements reflect true emotions of a person. Due to the short duration and low intensity, these micro-expressions are very difficult to perceive and interpret correctly. In this paper, we propose the dynamic representation of micro-expressions to preserve facial movement information of a video in a single frame. We also propose a Lateral Accretive Hybrid Network (LEARNet) to capture micro-level features of an expression in the facial region. The LEARNet refines the salient expression features in accretive manner by incorporating accretion layers (AL) in the network. The response of the AL holds the hybrid feature maps generated by prior laterally connected convolution layers. Moreover, LEARNet architecture incorporates the cross decoupled relationship between convolution layers which helps in preserving the tiny but influential facial muscle change information. The visual responses of the proposed LEARNet depict the effectiveness of the system by preserving both high- and micro-level edge features of facial expression. The effectiveness of the proposed LEARNet is evaluated on four benchmark datasets: CASME-I, CASME-II, CAS(ME)'2 and SMIC. The experimental results after investigation show a significant improvement of 4.03%, 1.90%, 1.79% and 2.82% as compared with ResNet on CASME-I, CASME-II, CAS(ME)'2 and SMIC datasets respectively.
与普遍的面部表情不同,微表情具有微妙的、不自主的肌肉运动,且本质上持续时间很短。这些微小的肌肉运动反映了一个人的真实情绪。由于持续时间短且强度低,这些微表情很难被正确察觉和解读。在本文中,我们提出微表情的动态表示方法,以在单帧中保留视频的面部运动信息。我们还提出了一种横向累加混合网络(LEARNet)来捕捉面部区域表情的微观特征。LEARNet通过在网络中加入累加层(AL)以累加的方式细化显著的表情特征。AL的响应保存了由先前横向连接的卷积层生成的混合特征图。此外,LEARNet架构纳入了卷积层之间的交叉解耦关系,这有助于保留微小但有影响的面部肌肉变化信息。所提出的LEARNet的视觉响应通过保留面部表情的高分辨率和微观边缘特征来体现系统的有效性。在四个基准数据集CASME-I、CASME-II、CAS(ME)'2和SMIC上评估了所提出的LEARNet的有效性。调查后的实验结果表明,与ResNet相比,在CASME-I、CASME-II、CAS(ME)'2和SMIC数据集上分别有4.03%、1.90%、1.79%和2.82%的显著提升。