Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild
The interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and ach...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Hindawi Limited
2021-01-01
|
Series: | Journal of Sensors |
Online Access: | http://dx.doi.org/10.1155/2021/8893661 |
id |
doaj-e9b32674ae6044c1a06c1214b2396491 |
---|---|
record_format |
Article |
spelling |
doaj-e9b32674ae6044c1a06c1214b23964912021-03-29T00:10:20ZengHindawi LimitedJournal of Sensors1687-72682021-01-01202110.1155/2021/8893661Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-WildAsad Ullah0Jing Wang1M. Shahid Anwar2Taeg Keun Whangbo3Yaping Zhu4Department of Computer Science and Information TechnologySchool of Information and ElectronicsSchool of Information and ElectronicsDepartment of Computer EngineeringSchool of Information and Communication EngineeringThe interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and achieved remarkable results; however, a significant improvement is required in spatial problems. This research work presents a novel framework and proposes an effective and robust solution for FER under an unconstrained environment. Face detection is performed using the supervision of facial attributes. Faceness-Net is used for deep facial part responses for the detection of faces under severe unconstrained variations. In order to improve the generalization problems and avoid insufficient data regime, Deep Convolutional Graphical Adversarial Network (DC-GAN) is utilized. Due to the challenging environmental factors faced in the wild, a large number of noises disrupt feature extraction, thus making it hard to capture ground truth. We leverage different multimodal sensors with a camera that aids in data acquisition, by extracting the features more accurately and improve the overall performance of FER. These intelligent sensors are used to tackle the significant challenges like illumination variance, subject dependence, and head pose. Dual-enhanced capsule network is used which is able to handle the spatial problem. The traditional capsule networks are unable to sufficiently extract the features, as the distance varies greatly between facial features. Therefore, the proposed network is capable of spatial transformation due to action unit aware mechanism and thus forward most desiring features for dynamic routing between capsules. Squashing function is used for the classification function. We have elaborated the effectiveness of our method by validating the results on four popular and versatile databases that outperform all state-of-the-art methods.http://dx.doi.org/10.1155/2021/8893661 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Asad Ullah Jing Wang M. Shahid Anwar Taeg Keun Whangbo Yaping Zhu |
spellingShingle |
Asad Ullah Jing Wang M. Shahid Anwar Taeg Keun Whangbo Yaping Zhu Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild Journal of Sensors |
author_facet |
Asad Ullah Jing Wang M. Shahid Anwar Taeg Keun Whangbo Yaping Zhu |
author_sort |
Asad Ullah |
title |
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild |
title_short |
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild |
title_full |
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild |
title_fullStr |
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild |
title_full_unstemmed |
Empirical Investigation of Multimodal Sensors in Novel Deep Facial Expression Recognition In-the-Wild |
title_sort |
empirical investigation of multimodal sensors in novel deep facial expression recognition in-the-wild |
publisher |
Hindawi Limited |
series |
Journal of Sensors |
issn |
1687-7268 |
publishDate |
2021-01-01 |
description |
The interest in the facial expression recognition (FER) is increasing day by day due to its practical and potential applications, such as human physiological interaction diagnosis and mental diseases detection. This area has received much attention from the research community in recent years and achieved remarkable results; however, a significant improvement is required in spatial problems. This research work presents a novel framework and proposes an effective and robust solution for FER under an unconstrained environment. Face detection is performed using the supervision of facial attributes. Faceness-Net is used for deep facial part responses for the detection of faces under severe unconstrained variations. In order to improve the generalization problems and avoid insufficient data regime, Deep Convolutional Graphical Adversarial Network (DC-GAN) is utilized. Due to the challenging environmental factors faced in the wild, a large number of noises disrupt feature extraction, thus making it hard to capture ground truth. We leverage different multimodal sensors with a camera that aids in data acquisition, by extracting the features more accurately and improve the overall performance of FER. These intelligent sensors are used to tackle the significant challenges like illumination variance, subject dependence, and head pose. Dual-enhanced capsule network is used which is able to handle the spatial problem. The traditional capsule networks are unable to sufficiently extract the features, as the distance varies greatly between facial features. Therefore, the proposed network is capable of spatial transformation due to action unit aware mechanism and thus forward most desiring features for dynamic routing between capsules. Squashing function is used for the classification function. We have elaborated the effectiveness of our method by validating the results on four popular and versatile databases that outperform all state-of-the-art methods. |
url |
http://dx.doi.org/10.1155/2021/8893661 |
work_keys_str_mv |
AT asadullah empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild AT jingwang empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild AT mshahidanwar empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild AT taegkeunwhangbo empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild AT yapingzhu empiricalinvestigationofmultimodalsensorsinnoveldeepfacialexpressionrecognitioninthewild |
_version_ |
1714761054942134272 |