Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions

Wearable motion sensors are assumed to be correctly positioned and oriented in most of the existing studies. However, generic wireless sensor units, patient health and state monitoring sensors, and smart phones and watches that contain sensors can be differently oriented on the body. The vast majori...

Full description

Bibliographic Details
Main Authors: Aras Yurtman, Billur Barshan, Barış Fidan
Format: Article
Language:English
Published: MDPI AG 2018-08-01
Series:Sensors
Subjects:
Online Access:http://www.mdpi.com/1424-8220/18/8/2725
id doaj-70371d7309ee49d79d72561f94145721
record_format Article
spelling doaj-70371d7309ee49d79d72561f941457212020-11-25T01:03:13ZengMDPI AGSensors1424-82202018-08-01188272510.3390/s18082725s18082725Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by QuaternionsAras Yurtman0Billur Barshan1Barış Fidan2Department of Electrical and Electronics Engineering, Bilkent University, Bilkent, Ankara 06800, TurkeyDepartment of Electrical and Electronics Engineering, Bilkent University, Bilkent, Ankara 06800, TurkeyDepartment of Mechanical and Mechatronics Engineering, University of Waterloo, 200 University Ave W, Waterloo, ON N2L 3G1, CanadaWearable motion sensors are assumed to be correctly positioned and oriented in most of the existing studies. However, generic wireless sensor units, patient health and state monitoring sensors, and smart phones and watches that contain sensors can be differently oriented on the body. The vast majority of the existing algorithms are not robust against placing the sensor units at variable orientations. We propose a method that transforms the recorded motion sensor sequences invariantly to sensor unit orientation. The method is based on estimating the sensor unit orientation and representing the sensor data with respect to the Earth frame. We also calculate the sensor rotations between consecutive time samples and represent them by quaternions in the Earth frame. We incorporate our method in the pre-processing stage of the standard activity recognition scheme and provide a comparative evaluation with the existing methods based on seven state-of-the-art classifiers and a publicly available dataset. The standard system with fixed sensor unit orientations cannot handle incorrectly oriented sensors, resulting in an average accuracy reduction of 31.8%. Our method results in an accuracy drop of only 4.7% on average compared to the standard system, outperforming the existing approaches that cause an accuracy degradation between 8.4 and 18.8%. We also consider stationary and non-stationary activities separately and evaluate the performance of each method for these two groups of activities. All of the methods perform significantly better in distinguishing non-stationary activities, our method resulting in an accuracy drop of 2.1% in this case. Our method clearly surpasses the remaining methods in classifying stationary activities where some of the methods noticeably fail. The proposed method is applicable to a wide range of wearable systems to make them robust against variable sensor unit orientations by transforming the sensor data at the pre-processing stage.http://www.mdpi.com/1424-8220/18/8/2725activity recognition and monitoringpatient health and state monitoringwearable sensingorientation-invariant sensingmotion sensorsaccelerometergyroscopemagnetometerpattern classification
collection DOAJ
language English
format Article
sources DOAJ
author Aras Yurtman
Billur Barshan
Barış Fidan
spellingShingle Aras Yurtman
Billur Barshan
Barış Fidan
Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
Sensors
activity recognition and monitoring
patient health and state monitoring
wearable sensing
orientation-invariant sensing
motion sensors
accelerometer
gyroscope
magnetometer
pattern classification
author_facet Aras Yurtman
Billur Barshan
Barış Fidan
author_sort Aras Yurtman
title Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
title_short Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
title_full Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
title_fullStr Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
title_full_unstemmed Activity Recognition Invariant to Wearable Sensor Unit Orientation Using Differential Rotational Transformations Represented by Quaternions
title_sort activity recognition invariant to wearable sensor unit orientation using differential rotational transformations represented by quaternions
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2018-08-01
description Wearable motion sensors are assumed to be correctly positioned and oriented in most of the existing studies. However, generic wireless sensor units, patient health and state monitoring sensors, and smart phones and watches that contain sensors can be differently oriented on the body. The vast majority of the existing algorithms are not robust against placing the sensor units at variable orientations. We propose a method that transforms the recorded motion sensor sequences invariantly to sensor unit orientation. The method is based on estimating the sensor unit orientation and representing the sensor data with respect to the Earth frame. We also calculate the sensor rotations between consecutive time samples and represent them by quaternions in the Earth frame. We incorporate our method in the pre-processing stage of the standard activity recognition scheme and provide a comparative evaluation with the existing methods based on seven state-of-the-art classifiers and a publicly available dataset. The standard system with fixed sensor unit orientations cannot handle incorrectly oriented sensors, resulting in an average accuracy reduction of 31.8%. Our method results in an accuracy drop of only 4.7% on average compared to the standard system, outperforming the existing approaches that cause an accuracy degradation between 8.4 and 18.8%. We also consider stationary and non-stationary activities separately and evaluate the performance of each method for these two groups of activities. All of the methods perform significantly better in distinguishing non-stationary activities, our method resulting in an accuracy drop of 2.1% in this case. Our method clearly surpasses the remaining methods in classifying stationary activities where some of the methods noticeably fail. The proposed method is applicable to a wide range of wearable systems to make them robust against variable sensor unit orientations by transforming the sensor data at the pre-processing stage.
topic activity recognition and monitoring
patient health and state monitoring
wearable sensing
orientation-invariant sensing
motion sensors
accelerometer
gyroscope
magnetometer
pattern classification
url http://www.mdpi.com/1424-8220/18/8/2725
work_keys_str_mv AT arasyurtman activityrecognitioninvarianttowearablesensorunitorientationusingdifferentialrotationaltransformationsrepresentedbyquaternions
AT billurbarshan activityrecognitioninvarianttowearablesensorunitorientationusingdifferentialrotationaltransformationsrepresentedbyquaternions
AT barısfidan activityrecognitioninvarianttowearablesensorunitorientationusingdifferentialrotationaltransformationsrepresentedbyquaternions
_version_ 1725201713882726400