Human action recognition based on low- and high-level data from wearable inertial sensors
Human action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertia...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
SAGE Publishing
2019-12-01
|
Series: | International Journal of Distributed Sensor Networks |
Online Access: | https://doi.org/10.1177/1550147719894532 |
id |
doaj-559fe9c3e5d845d9ad0bb94c052b8bc5 |
---|---|
record_format |
Article |
spelling |
doaj-559fe9c3e5d845d9ad0bb94c052b8bc52020-11-25T03:49:38ZengSAGE PublishingInternational Journal of Distributed Sensor Networks1550-14772019-12-011510.1177/1550147719894532Human action recognition based on low- and high-level data from wearable inertial sensorsIrvin Hussein Lopez-Nava0Angélica Muñoz-Meléndez1Instituto Nacional de Astrofísica, Óptica y Electrónica, San Andrés Cholula, MexicoInstituto Nacional de Astrofísica, Óptica y Electrónica, San Andrés Cholula, MexicoHuman action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertial sensors, in which raw sensor data are used to build classification models, and in a few of them high-level representations are obtained which are directly related to anatomical characteristics of the human body. This research focuses on classifying a set of activities of daily living, such as functional mobility, and instrumental activities of daily living, such as preparing meals, performed by test subjects in their homes in naturalistic conditions. The joint angles of upper and lower limbs are estimated using information from five wearable inertial sensors placed on the body of five test subjects. A set of features related to human limb motions is extracted from the orientation signals (high-level data) and another set from the acceleration raw signals (low-level data) and both are used to build classifiers using four inference algorithms. The proposed features in this work are the number of movements and the average duration of consecutive movements. The classifiers are capable of successfully classifying the set of actions using raw data with up to 77.8% and 93.3% from high-level data. This study allowed comparing the use of two data levels to classify a set of actions performed in daily environments using an inertial sensor network.https://doi.org/10.1177/1550147719894532 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Irvin Hussein Lopez-Nava Angélica Muñoz-Meléndez |
spellingShingle |
Irvin Hussein Lopez-Nava Angélica Muñoz-Meléndez Human action recognition based on low- and high-level data from wearable inertial sensors International Journal of Distributed Sensor Networks |
author_facet |
Irvin Hussein Lopez-Nava Angélica Muñoz-Meléndez |
author_sort |
Irvin Hussein Lopez-Nava |
title |
Human action recognition based on low- and high-level data from wearable inertial sensors |
title_short |
Human action recognition based on low- and high-level data from wearable inertial sensors |
title_full |
Human action recognition based on low- and high-level data from wearable inertial sensors |
title_fullStr |
Human action recognition based on low- and high-level data from wearable inertial sensors |
title_full_unstemmed |
Human action recognition based on low- and high-level data from wearable inertial sensors |
title_sort |
human action recognition based on low- and high-level data from wearable inertial sensors |
publisher |
SAGE Publishing |
series |
International Journal of Distributed Sensor Networks |
issn |
1550-1477 |
publishDate |
2019-12-01 |
description |
Human action recognition supported by highly accurate specialized systems, ambulatory systems, or wireless sensor networks has a tremendous potential in the areas of healthcare or wellbeing monitoring. Recently, several studies carried out focused on the recognition of actions using wearable inertial sensors, in which raw sensor data are used to build classification models, and in a few of them high-level representations are obtained which are directly related to anatomical characteristics of the human body. This research focuses on classifying a set of activities of daily living, such as functional mobility, and instrumental activities of daily living, such as preparing meals, performed by test subjects in their homes in naturalistic conditions. The joint angles of upper and lower limbs are estimated using information from five wearable inertial sensors placed on the body of five test subjects. A set of features related to human limb motions is extracted from the orientation signals (high-level data) and another set from the acceleration raw signals (low-level data) and both are used to build classifiers using four inference algorithms. The proposed features in this work are the number of movements and the average duration of consecutive movements. The classifiers are capable of successfully classifying the set of actions using raw data with up to 77.8% and 93.3% from high-level data. This study allowed comparing the use of two data levels to classify a set of actions performed in daily environments using an inertial sensor network. |
url |
https://doi.org/10.1177/1550147719894532 |
work_keys_str_mv |
AT irvinhusseinlopeznava humanactionrecognitionbasedonlowandhighleveldatafromwearableinertialsensors AT angelicamunozmelendez humanactionrecognitionbasedonlowandhighleveldatafromwearableinertialsensors |
_version_ |
1724494259326812160 |