Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication

During social interaction, humans recognize others’ emotions via individual features and interpersonal features. However, most previous automatic emotion recognition techniques only used individual features—they have not tested the importance of interpersonal features. In the present study, we asked...

Full description

Bibliographic Details
Main Authors: Jingyu Quan, Yoshihiro Miyake, Takayuki Nozawa
Format: Article
Language:English
Published: MDPI AG 2021-08-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/16/5317
id doaj-ca429b4a68ac426a85bd066d000af200
record_format Article
spelling doaj-ca429b4a68ac426a85bd066d000af2002021-08-26T14:18:38ZengMDPI AGSensors1424-82202021-08-01215317531710.3390/s21165317Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during CommunicationJingyu Quan0Yoshihiro Miyake1Takayuki Nozawa2Department of Computer Science, Tokyo Institute of Technology, Yokohama 226-8502, JapanDepartment of Computer Science, Tokyo Institute of Technology, Yokohama 226-8502, JapanResearch Institute for the Earth Inclusive Sensing, Tokyo Institute of Technology, Tokyo 152-8550, JapanDuring social interaction, humans recognize others’ emotions via individual features and interpersonal features. However, most previous automatic emotion recognition techniques only used individual features—they have not tested the importance of interpersonal features. In the present study, we asked whether interpersonal features, especially time-lagged synchronization features, are beneficial to the performance of automatic emotion recognition techniques. We explored this question in the main experiment (speaker-dependent emotion recognition) and supplementary experiment (speaker-independent emotion recognition) by building an individual framework and interpersonal framework in visual, audio, and cross-modality, respectively. Our main experiment results showed that the interpersonal framework outperformed the individual framework in every modality. Our supplementary experiment showed—even for unknown communication pairs—that the interpersonal framework led to a better performance. Therefore, we concluded that interpersonal features are useful to boost the performance of automatic emotion recognition tasks. We hope to raise attention to interpersonal features in this study.https://www.mdpi.com/1424-8220/21/16/5317affective computingclassificationcommunicationdeep neural networksemotion recognitioninterpersonal features
collection DOAJ
language English
format Article
sources DOAJ
author Jingyu Quan
Yoshihiro Miyake
Takayuki Nozawa
spellingShingle Jingyu Quan
Yoshihiro Miyake
Takayuki Nozawa
Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
Sensors
affective computing
classification
communication
deep neural networks
emotion recognition
interpersonal features
author_facet Jingyu Quan
Yoshihiro Miyake
Takayuki Nozawa
author_sort Jingyu Quan
title Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
title_short Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
title_full Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
title_fullStr Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
title_full_unstemmed Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
title_sort incorporating interpersonal synchronization features for automatic emotion recognition from visual and audio data during communication
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2021-08-01
description During social interaction, humans recognize others’ emotions via individual features and interpersonal features. However, most previous automatic emotion recognition techniques only used individual features—they have not tested the importance of interpersonal features. In the present study, we asked whether interpersonal features, especially time-lagged synchronization features, are beneficial to the performance of automatic emotion recognition techniques. We explored this question in the main experiment (speaker-dependent emotion recognition) and supplementary experiment (speaker-independent emotion recognition) by building an individual framework and interpersonal framework in visual, audio, and cross-modality, respectively. Our main experiment results showed that the interpersonal framework outperformed the individual framework in every modality. Our supplementary experiment showed—even for unknown communication pairs—that the interpersonal framework led to a better performance. Therefore, we concluded that interpersonal features are useful to boost the performance of automatic emotion recognition tasks. We hope to raise attention to interpersonal features in this study.
topic affective computing
classification
communication
deep neural networks
emotion recognition
interpersonal features
url https://www.mdpi.com/1424-8220/21/16/5317
work_keys_str_mv AT jingyuquan incorporatinginterpersonalsynchronizationfeaturesforautomaticemotionrecognitionfromvisualandaudiodataduringcommunication
AT yoshihiromiyake incorporatinginterpersonalsynchronizationfeaturesforautomaticemotionrecognitionfromvisualandaudiodataduringcommunication
AT takayukinozawa incorporatinginterpersonalsynchronizationfeaturesforautomaticemotionrecognitionfromvisualandaudiodataduringcommunication
_version_ 1721190201726861312