TY - GEN
T1 - Analyzing facial temporal patterns for face anti-spoofing
AU - Xia, Jingtian
AU - Song, Siyang
AU - Tang, Yan
AU - Shen, Linlin
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/10/30
Y1 - 2020/10/30
N2 - Face anti-spoofing is crucial as face recognition systems are widely challenged by the print attack and replay attack. Since facial temporal patterns of these attacks and real face are naturally different, this paper proposes two temporal modelling approaches to face anti-spoofing tasks. Firstly, we propose to analyze the temporal patterns of mid-level facial attributes in spectral domain, aiming to find the unique frequency patterns of real face and each attack, respectively. Then, we propose to directly model dynamics from the given data, by employing the dynamic image algorithm to generate low-level spatiotemporal representations of videos. In particular, we extract deep features from both global and local face parts, i.e. eyes, nose and mouth, and then fuse them for face spoofing detection. Then, a Convolutional Neural Networks (CNN) - Long Short-Term Memory (LSTM) units (CNN-LSTM) architecture is introduced to learn the high-level spatiotemporal features from dynamic facial images. The proposed approaches were evaluated on two benchmark databases. The results suggest the effectiveness of the second approaches, i.e. as low as 1.85% Equal Error Rate (EER) on CASIA-FASD and 0.00% Average Classification Error Rate (ACER) on REPLAY-ATTACK have been achieved.
AB - Face anti-spoofing is crucial as face recognition systems are widely challenged by the print attack and replay attack. Since facial temporal patterns of these attacks and real face are naturally different, this paper proposes two temporal modelling approaches to face anti-spoofing tasks. Firstly, we propose to analyze the temporal patterns of mid-level facial attributes in spectral domain, aiming to find the unique frequency patterns of real face and each attack, respectively. Then, we propose to directly model dynamics from the given data, by employing the dynamic image algorithm to generate low-level spatiotemporal representations of videos. In particular, we extract deep features from both global and local face parts, i.e. eyes, nose and mouth, and then fuse them for face spoofing detection. Then, a Convolutional Neural Networks (CNN) - Long Short-Term Memory (LSTM) units (CNN-LSTM) architecture is introduced to learn the high-level spatiotemporal features from dynamic facial images. The proposed approaches were evaluated on two benchmark databases. The results suggest the effectiveness of the second approaches, i.e. as low as 1.85% Equal Error Rate (EER) on CASIA-FASD and 0.00% Average Classification Error Rate (ACER) on REPLAY-ATTACK have been achieved.
KW - Convolutional neural networks
KW - Dynamic image
KW - Face anti-spoofing
KW - Fourier transform
KW - Long short-term memory
UR - http://www.scopus.com/inward/record.url?scp=85099877924&partnerID=8YFLogxK
U2 - 10.1145/3436369.3437404
DO - 10.1145/3436369.3437404
M3 - Conference contribution
AN - SCOPUS:85099877924
T3 - ACM International Conference Proceeding Series
SP - 200
EP - 207
BT - ICCPR 2020 - Proceedings of 2020 9th International Conference on Computing and Pattern Recognition
PB - Association for Computing Machinery
T2 - 9th International Conference on Computing and Pattern Recognition, ICCPR 2020
Y2 - 30 October 2020 through 1 November 2020
ER -