Human action recognition based on low- and high-level data from wearable inertial sensors

Human action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertia...

Full description

Bibliographic Details
Main Authors: Irvin Hussein Lopez-Nava, Angélica Muñoz-Meléndez
Format: Article
Language:English
Published: SAGE Publishing 2019-12-01
Series:International Journal of Distributed Sensor Networks
Online Access:https://doi.org/10.1177/1550147719894532
id doaj-559fe9c3e5d845d9ad0bb94c052b8bc5
record_format Article
spelling doaj-559fe9c3e5d845d9ad0bb94c052b8bc52020-11-25T03:49:38ZengSAGE PublishingInternational Journal of Distributed Sensor Networks1550-14772019-12-011510.1177/1550147719894532Human action recognition based on low- and high-level data from wearable inertial sensorsIrvin Hussein Lopez-Nava0Angélica Muñoz-Meléndez1Instituto Nacional de Astrofísica, Óptica y Electrónica, San Andrés Cholula, MexicoInstituto Nacional de Astrofísica, Óptica y Electrónica, San Andrés Cholula, MexicoHuman action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertial sensors, in which raw sensor data are used to build classification models, and in a few of them high-level representations are obtained which are directly related to anatomical characteristics of the human body. This research focuses on classifying a set of activities of daily living, such as functional mobility, and instrumental activities of daily living, such as preparing meals, performed by test subjects in their homes in naturalistic conditions. The joint angles of upper and lower limbs are estimated using information from five wearable inertial sensors placed on the body of five test subjects. A set of features related to human limb motions is extracted from the orientation signals (high-level data) and another set from the acceleration raw signals (low-level data) and both are used to build classifiers using four inference algorithms. The proposed features in this work are the number of movements and the average duration of consecutive movements. The classifiers are capable of successfully classifying the set of actions using raw data with up to 77.8% and 93.3% from high-level data. This study allowed comparing the use of two data levels to classify a set of actions performed in daily environments using an inertial sensor network.https://doi.org/10.1177/1550147719894532
collection DOAJ
language English
format Article
sources DOAJ
author Irvin Hussein Lopez-Nava
Angélica Muñoz-Meléndez
spellingShingle Irvin Hussein Lopez-Nava
Angélica Muñoz-Meléndez
Human action recognition based on low- and high-level data from wearable inertial sensors
International Journal of Distributed Sensor Networks
author_facet Irvin Hussein Lopez-Nava
Angélica Muñoz-Meléndez
author_sort Irvin Hussein Lopez-Nava
title Human action recognition based on low- and high-level data from wearable inertial sensors
title_short Human action recognition based on low- and high-level data from wearable inertial sensors
title_full Human action recognition based on low- and high-level data from wearable inertial sensors
title_fullStr Human action recognition based on low- and high-level data from wearable inertial sensors
title_full_unstemmed Human action recognition based on low- and high-level data from wearable inertial sensors
title_sort human action recognition based on low- and high-level data from wearable inertial sensors
publisher SAGE Publishing
series International Journal of Distributed Sensor Networks
issn 1550-1477
publishDate 2019-12-01
description Human action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertial sensors, in which raw sensor data are used to build classification models, and in a few of them high-level representations are obtained which are directly related to anatomical characteristics of the human body. This research focuses on classifying a set of activities of daily living, such as functional mobility, and instrumental activities of daily living, such as preparing meals, performed by test subjects in their homes in naturalistic conditions. The joint angles of upper and lower limbs are estimated using information from five wearable inertial sensors placed on the body of five test subjects. A set of features related to human limb motions is extracted from the orientation signals (high-level data) and another set from the acceleration raw signals (low-level data) and both are used to build classifiers using four inference algorithms. The proposed features in this work are the number of movements and the average duration of consecutive movements. The classifiers are capable of successfully classifying the set of actions using raw data with up to 77.8% and 93.3% from high-level data. This study allowed comparing the use of two data levels to classify a set of actions performed in daily environments using an inertial sensor network.
url https://doi.org/10.1177/1550147719894532
work_keys_str_mv AT irvinhusseinlopeznava humanactionrecognitionbasedonlowandhighleveldatafromwearableinertialsensors
AT angelicamunozmelendez humanactionrecognitionbasedonlowandhighleveldatafromwearableinertialsensors
_version_ 1724494259326812160