Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences

Face liveness detection is a critical preprocessing step in face recognition for avoiding face spoofing attacks, where an impostor can impersonate a valid user for authentication. While considerable research has been recently done in improving the accuracy of face liveness detection, the best curren...

Full description

Bibliographic Details
Main Authors: Ranjana Koshy, Ausif Mahmood
Format: Article
Language:English
Published: MDPI AG 2020-10-01
Series:Entropy
Subjects:
Online Access:https://www.mdpi.com/1099-4300/22/10/1186
id doaj-44d2cf362f85473fba54065b573ba21e
record_format Article
spelling doaj-44d2cf362f85473fba54065b573ba21e2020-11-25T01:43:13ZengMDPI AGEntropy1099-43002020-10-01221186118610.3390/e22101186Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video SequencesRanjana Koshy0Ausif Mahmood1Computer Science and Engineering Department, University of Bridgeport, Bridgeport, CT 06604, USAComputer Science and Engineering Department, University of Bridgeport, Bridgeport, CT 06604, USAFace liveness detection is a critical preprocessing step in face recognition for avoiding face spoofing attacks, where an impostor can impersonate a valid user for authentication. While considerable research has been recently done in improving the accuracy of face liveness detection, the best current approaches use a two-step process of first applying non-linear anisotropic diffusion to the incoming image and then using a deep network for final liveness decision. Such an approach is not viable for real-time face liveness detection. We develop two end-to-end real-time solutions where nonlinear anisotropic diffusion based on an additive operator splitting scheme is first applied to an incoming static image, which enhances the edges and surface texture, and preserves the boundary locations in the real image. The diffused image is then forwarded to a pre-trained Specialized Convolutional Neural Network (SCNN) and the Inception network version 4, which identify the complex and deep features for face liveness classification. We evaluate the performance of our integrated approach using the SCNN and Inception v4 on the Replay-Attack dataset and Replay-Mobile dataset. The entire architecture is created in such a manner that, once trained, the face liveness detection can be accomplished in real-time. We achieve promising results of 96.03% and 96.21% face liveness detection accuracy with the SCNN, and 94.77% and 95.53% accuracy with the Inception v4, on the Replay-Attack, and Replay-Mobile datasets, respectively. We also develop a novel deep architecture for face liveness detection on video frames that uses the diffusion of images followed by a deep Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) to classify the video sequence as real or fake. Even though the use of CNN followed by LSTM is not new, combining it with diffusion (that has proven to be the best approach for single image liveness detection) is novel. Performance evaluation of our architecture on the REPLAY-ATTACK dataset gave 98.71% test accuracy and 2.77% Half Total Error Rate (HTER), and on the REPLAY-MOBILE dataset gave 95.41% accuracy and 5.28% HTER.https://www.mdpi.com/1099-4300/22/10/1186face liveness detectiondiffusionSCNNInception v4CNN-LSTMReplay-Attack dataset
collection DOAJ
language English
format Article
sources DOAJ
author Ranjana Koshy
Ausif Mahmood
spellingShingle Ranjana Koshy
Ausif Mahmood
Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
Entropy
face liveness detection
diffusion
SCNN
Inception v4
CNN-LSTM
Replay-Attack dataset
author_facet Ranjana Koshy
Ausif Mahmood
author_sort Ranjana Koshy
title Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
title_short Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
title_full Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
title_fullStr Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
title_full_unstemmed Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences
title_sort enhanced deep learning architectures for face liveness detection for static and video sequences
publisher MDPI AG
series Entropy
issn 1099-4300
publishDate 2020-10-01
description Face liveness detection is a critical preprocessing step in face recognition for avoiding face spoofing attacks, where an impostor can impersonate a valid user for authentication. While considerable research has been recently done in improving the accuracy of face liveness detection, the best current approaches use a two-step process of first applying non-linear anisotropic diffusion to the incoming image and then using a deep network for final liveness decision. Such an approach is not viable for real-time face liveness detection. We develop two end-to-end real-time solutions where nonlinear anisotropic diffusion based on an additive operator splitting scheme is first applied to an incoming static image, which enhances the edges and surface texture, and preserves the boundary locations in the real image. The diffused image is then forwarded to a pre-trained Specialized Convolutional Neural Network (SCNN) and the Inception network version 4, which identify the complex and deep features for face liveness classification. We evaluate the performance of our integrated approach using the SCNN and Inception v4 on the Replay-Attack dataset and Replay-Mobile dataset. The entire architecture is created in such a manner that, once trained, the face liveness detection can be accomplished in real-time. We achieve promising results of 96.03% and 96.21% face liveness detection accuracy with the SCNN, and 94.77% and 95.53% accuracy with the Inception v4, on the Replay-Attack, and Replay-Mobile datasets, respectively. We also develop a novel deep architecture for face liveness detection on video frames that uses the diffusion of images followed by a deep Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) to classify the video sequence as real or fake. Even though the use of CNN followed by LSTM is not new, combining it with diffusion (that has proven to be the best approach for single image liveness detection) is novel. Performance evaluation of our architecture on the REPLAY-ATTACK dataset gave 98.71% test accuracy and 2.77% Half Total Error Rate (HTER), and on the REPLAY-MOBILE dataset gave 95.41% accuracy and 5.28% HTER.
topic face liveness detection
diffusion
SCNN
Inception v4
CNN-LSTM
Replay-Attack dataset
url https://www.mdpi.com/1099-4300/22/10/1186
work_keys_str_mv AT ranjanakoshy enhanceddeeplearningarchitecturesforfacelivenessdetectionforstaticandvideosequences
AT ausifmahmood enhanceddeeplearningarchitecturesforfacelivenessdetectionforstaticandvideosequences
_version_ 1725032661418770432