Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild

The interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and ach...

Full description

Bibliographic Details
Main Authors: Asad Ullah, Jing Wang, M. Shahid Anwar, Taeg Keun Whangbo, Yaping Zhu
Format: Article
Language:English
Published: Hindawi Limited 2021-01-01
Series:Journal of Sensors
Online Access:http://dx.doi.org/10.1155/2021/8893661
id doaj-e9b32674ae6044c1a06c1214b2396491
record_format Article
spelling doaj-e9b32674ae6044c1a06c1214b23964912021-03-29T00:10:20ZengHindawi LimitedJournal of Sensors1687-72682021-01-01202110.1155/2021/8893661Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-WildAsad Ullah0Jing Wang1M. Shahid Anwar2Taeg Keun Whangbo3Yaping Zhu4Department of Computer Science and Information TechnologySchool of Information and ElectronicsSchool of Information and ElectronicsDepartment of Computer EngineeringSchool of Information and Communication EngineeringThe interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and achieved remarkable results; however, a significant improvement is required in spatial problems. This research work presents a novel framework and proposes an effective and robust solution for FER under an unconstrained environment. Face detection is performed using the supervision of facial attributes. Faceness-Net is used for deep facial part responses for the detection of faces under severe unconstrained variations. In order to improve the generalization problems and avoid insufficient data regime, Deep Convolutional Graphical Adversarial Network (DC-GAN) is utilized. Due to the challenging environmental factors faced in the wild, a large number of noises disrupt feature extraction, thus making it hard to capture ground truth. We leverage different multimodal sensors with a camera that aids in data acquisition, by extracting the features more accurately and improve the overall performance of FER. These intelligent sensors are used to tackle the significant challenges like illumination variance, subject dependence, and head pose. Dual-enhanced capsule network is used which is able to handle the spatial problem. The traditional capsule networks are unable to sufficiently extract the features, as the distance varies greatly between facial features. Therefore, the proposed network is capable of spatial transformation due to action unit aware mechanism and thus forward most desiring features for dynamic routing between capsules. Squashing function is used for the classification function. We have elaborated the effectiveness of our method by validating the results on four popular and versatile databases that outperform all state-of-the-art methods.http://dx.doi.org/10.1155/2021/8893661
collection DOAJ
language English
format Article
sources DOAJ
author Asad Ullah
Jing Wang
M. Shahid Anwar
Taeg Keun Whangbo
Yaping Zhu
spellingShingle Asad Ullah
Jing Wang
M. Shahid Anwar
Taeg Keun Whangbo
Yaping Zhu
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
Journal of Sensors
author_facet Asad Ullah
Jing Wang
M. Shahid Anwar
Taeg Keun Whangbo
Yaping Zhu
author_sort Asad Ullah
title Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
title_short Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
title_full Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
title_fullStr Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
title_full_unstemmed Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
title_sort empirical investigation of multimodal sensors in novel deep facial expression recognition in-the-wild
publisher Hindawi Limited
series Journal of Sensors
issn 1687-7268
publishDate 2021-01-01
description The interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and achieved remarkable results; however, a significant improvement is required in spatial problems. This research work presents a novel framework and proposes an effective and robust solution for FER under an unconstrained environment. Face detection is performed using the supervision of facial attributes. Faceness-Net is used for deep facial part responses for the detection of faces under severe unconstrained variations. In order to improve the generalization problems and avoid insufficient data regime, Deep Convolutional Graphical Adversarial Network (DC-GAN) is utilized. Due to the challenging environmental factors faced in the wild, a large number of noises disrupt feature extraction, thus making it hard to capture ground truth. We leverage different multimodal sensors with a camera that aids in data acquisition, by extracting the features more accurately and improve the overall performance of FER. These intelligent sensors are used to tackle the significant challenges like illumination variance, subject dependence, and head pose. Dual-enhanced capsule network is used which is able to handle the spatial problem. The traditional capsule networks are unable to sufficiently extract the features, as the distance varies greatly between facial features. Therefore, the proposed network is capable of spatial transformation due to action unit aware mechanism and thus forward most desiring features for dynamic routing between capsules. Squashing function is used for the classification function. We have elaborated the effectiveness of our method by validating the results on four popular and versatile databases that outperform all state-of-the-art methods.
url http://dx.doi.org/10.1155/2021/8893661
work_keys_str_mv AT asadullah empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild
AT jingwang empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild
AT mshahidanwar empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild
AT taegkeunwhangbo empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild
AT yapingzhu empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild
_version_ 1714761054942134272