Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models

Automatic emotion recognition has been widely studied and applied to various computer vision tasks (e.g. health monitoring, driver state surveillance, personalized learning, and security monitoring). With the great potential provided by current advanced 3D scanners technology (e.g. the Kinect), we s...

Full description

Bibliographic Details
Main Author: Zhang, Yang
Other Authors: Zhang, Li
Published: Northumbria University 2015
Subjects:
Online Access:http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.664670
id ndltd-bl.uk-oai-ethos.bl.uk-664670
record_format oai_dc
spelling ndltd-bl.uk-oai-ethos.bl.uk-6646702017-04-20T03:17:50ZIntelligent emotion recognition from facial and whole-body expressions using adaptive ensemble modelsZhang, YangZhang, Li2015Automatic emotion recognition has been widely studied and applied to various computer vision tasks (e.g. health monitoring, driver state surveillance, personalized learning, and security monitoring). With the great potential provided by current advanced 3D scanners technology (e.g. the Kinect), we shed light on robust emotion recognition based one users’ facial and whole-body expressions. As revealed by recent psychological and behavioral research, facial expressions are good in communicating categorical emotions (e.g. happy, sad, surprise, etc.), while bodily expressions could contribute more to the perception of dimensional emotional states (e.g. the arousal and valence dimensions). Thus, we propose two novel emotion recognition systems respectively applying adaptive ensemble classification and regression models respectively based on the facial and bodily modalities. The proposed real-time 3D facial Action Unit (AU) intensity estimation and emotion recognition system automatically selects 16 motion-based facial feature sets to estimate the intensities of 16 diagnostic AUs. Then a set of six novel adaptive ensemble classifiers are proposed for robust classification of the six basic emotions and the detection of newly arrived unseen novel emotion classes (emotions that are not included in the training set). In both offline-line and on-line real-time evaluation, the system shows the highest recognition accuracy in comparison with other related work and flexibility and good adaptation for newly arrived novel emotion detection(e.g. ‘contempt’ which is not included in the six basic emotions). The second system focuses on continuous and dimensional affect prediction from users’ bodily expressions using adaptive regression. Both static posture and dynamic motion bodily features are extracted and subsequently selected by a Genetic Algorithm to identify their most discriminative combinations for both valence and arousal dimensions. Then an adaptive ensemble regression model is proposed to robustly map subjects’ emotional states onto a continuous arousal-valence affective space using the identified feature subsets. Experimental results show that the proposed system outperforms other benchmark models and achieves promising performance compared to other state-of-the-art research reported in the literature. Furthermore, we also propose a novel semi-feature level bimodal fusion framework that integrates both facial and bodily information together to draw a more comprehensive and robust dimensional interpretation of subjects’ emotional states. By combining the optimal discriminative bodily features and the derived AU intensities as inputs, the proposed adaptive ensemble regression model achieves remarkable improvements in comparison to solely applying the bodily features.006.3G400 Computer ScienceNorthumbria Universityhttp://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.664670http://nrl.northumbria.ac.uk/23588/Electronic Thesis or Dissertation
collection NDLTD
sources NDLTD
topic 006.3
G400 Computer Science
spellingShingle 006.3
G400 Computer Science
Zhang, Yang
Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
description Automatic emotion recognition has been widely studied and applied to various computer vision tasks (e.g. health monitoring, driver state surveillance, personalized learning, and security monitoring). With the great potential provided by current advanced 3D scanners technology (e.g. the Kinect), we shed light on robust emotion recognition based one users’ facial and whole-body expressions. As revealed by recent psychological and behavioral research, facial expressions are good in communicating categorical emotions (e.g. happy, sad, surprise, etc.), while bodily expressions could contribute more to the perception of dimensional emotional states (e.g. the arousal and valence dimensions). Thus, we propose two novel emotion recognition systems respectively applying adaptive ensemble classification and regression models respectively based on the facial and bodily modalities. The proposed real-time 3D facial Action Unit (AU) intensity estimation and emotion recognition system automatically selects 16 motion-based facial feature sets to estimate the intensities of 16 diagnostic AUs. Then a set of six novel adaptive ensemble classifiers are proposed for robust classification of the six basic emotions and the detection of newly arrived unseen novel emotion classes (emotions that are not included in the training set). In both offline-line and on-line real-time evaluation, the system shows the highest recognition accuracy in comparison with other related work and flexibility and good adaptation for newly arrived novel emotion detection(e.g. ‘contempt’ which is not included in the six basic emotions). The second system focuses on continuous and dimensional affect prediction from users’ bodily expressions using adaptive regression. Both static posture and dynamic motion bodily features are extracted and subsequently selected by a Genetic Algorithm to identify their most discriminative combinations for both valence and arousal dimensions. Then an adaptive ensemble regression model is proposed to robustly map subjects’ emotional states onto a continuous arousal-valence affective space using the identified feature subsets. Experimental results show that the proposed system outperforms other benchmark models and achieves promising performance compared to other state-of-the-art research reported in the literature. Furthermore, we also propose a novel semi-feature level bimodal fusion framework that integrates both facial and bodily information together to draw a more comprehensive and robust dimensional interpretation of subjects’ emotional states. By combining the optimal discriminative bodily features and the derived AU intensities as inputs, the proposed adaptive ensemble regression model achieves remarkable improvements in comparison to solely applying the bodily features.
author2 Zhang, Li
author_facet Zhang, Li
Zhang, Yang
author Zhang, Yang
author_sort Zhang, Yang
title Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
title_short Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
title_full Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
title_fullStr Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
title_full_unstemmed Intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
title_sort intelligent emotion recognition from facial and whole-body expressions using adaptive ensemble models
publisher Northumbria University
publishDate 2015
url http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.664670
work_keys_str_mv AT zhangyang intelligentemotionrecognitionfromfacialandwholebodyexpressionsusingadaptiveensemblemodels
_version_ 1718439777438531584