School of Engineering, University of Kent, Canterbury CT2 7NT, UK.
Sensors (Basel). 2022 Apr 28;22(9):3365. doi: 10.3390/s22093365.
Despite the high performances achieved using deep learning techniques in biometric systems, the inability to rationalise the decisions reached by such approaches is a significant drawback for the usability and security requirements of many applications. For Facial Biometric Presentation Attack Detection (PAD), deep learning approaches can provide good classification results but cannot answer the questions such as "Why did the system make this decision"? To overcome this limitation, an explainable deep neural architecture for Facial Biometric Presentation Attack Detection is introduced in this paper. Both visual and verbal explanations are produced using the saliency maps from a Grad-CAM approach and the gradient from a Long-Short-Term-Memory (LSTM) network with a modified gate function. These explanations have also been used in the proposed framework as additional information to further improve the classification performance. The proposed framework utilises both spatial and temporal information to help the model focus on anomalous visual characteristics that indicate spoofing attacks. The performance of the proposed approach is evaluated using the CASIA-FA, Replay Attack, MSU-MFSD, and HKBU MARs datasets and indicates the effectiveness of the proposed method for improving performance and producing usable explanations.
尽管深度学习技术在生物识别系统中取得了很高的性能,但这种方法无法合理说明其决策,这对于许多应用的可用性和安全性要求来说是一个重大的缺点。对于面部生物特征呈现攻击检测 (PAD),深度学习方法可以提供很好的分类结果,但无法回答“系统为什么做出这个决定”等问题。为了克服这一限制,本文引入了一种可解释的深度学习架构用于面部生物特征呈现攻击检测。通过 Grad-CAM 方法的显着性图和具有修改后的门控函数的长短期记忆 (LSTM) 网络的梯度来生成视觉和口头解释。这些解释也被用作附加信息用于进一步提高分类性能。所提出的框架利用空间和时间信息来帮助模型关注异常的视觉特征,这些特征表明存在欺骗攻击。所提出的方法的性能使用 CASIA-FA、Replay Attack、MSU-MFSD 和 HKBU MARs 数据集进行评估,并表明该方法在提高性能和生成可用解释方面的有效性。