Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.

Real-time human activity recognition is essential for human-robot interactions for assisted healthy independent living. Most previous work in this area is performed on traditional two-dimensional (2D) videos and both global and local methods have been used. Since 2D videos are sensitive to changes o...

Full description

Bibliographic Details
Main Authors: Jin Qi, Zhiyong Yang
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2014-01-01
Series:PLoS ONE
Online Access:http://europepmc.org/articles/PMC4256388?pdf=render
id doaj-027c0cb9510d4693bca3de514aa21fed
record_format Article
spelling doaj-027c0cb9510d4693bca3de514aa21fed2020-11-25T01:27:43ZengPublic Library of Science (PLoS)PLoS ONE1932-62032014-01-01912e11414710.1371/journal.pone.0114147Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.Jin QiZhiyong YangReal-time human activity recognition is essential for human-robot interactions for assisted healthy independent living. Most previous work in this area is performed on traditional two-dimensional (2D) videos and both global and local methods have been used. Since 2D videos are sensitive to changes of lighting condition, view angle, and scale, researchers begun to explore applications of 3D information in human activity understanding in recently years. Unfortunately, features that work well on 2D videos usually don't perform well on 3D videos and there is no consensus on what 3D features should be used. Here we propose a model of human activity recognition based on 3D movements of body joints. Our method has three steps, learning dictionaries of sparse codes of 3D movements of joints, sparse coding, and classification. In the first step, space-time volumes of 3D movements of body joints are obtained via dense sampling and independent component analysis is then performed to construct a dictionary of sparse codes for each activity. In the second step, the space-time volumes are projected to the dictionaries and a set of sparse histograms of the projection coefficients are constructed as feature representations of the activities. Finally, the sparse histograms are used as inputs to a support vector machine to recognize human activities. We tested this model on three databases of human activities and found that it outperforms the state-of-the-art algorithms. Thus, this model can be used for real-time human activity recognition in many applications.http://europepmc.org/articles/PMC4256388?pdf=render
collection DOAJ
language English
format Article
sources DOAJ
author Jin Qi
Zhiyong Yang
spellingShingle Jin Qi
Zhiyong Yang
Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
PLoS ONE
author_facet Jin Qi
Zhiyong Yang
author_sort Jin Qi
title Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
title_short Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
title_full Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
title_fullStr Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
title_full_unstemmed Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.
title_sort learning dictionaries of sparse codes of 3d movements of body joints for real-time human activity understanding.
publisher Public Library of Science (PLoS)
series PLoS ONE
issn 1932-6203
publishDate 2014-01-01
description Real-time human activity recognition is essential for human-robot interactions for assisted healthy independent living. Most previous work in this area is performed on traditional two-dimensional (2D) videos and both global and local methods have been used. Since 2D videos are sensitive to changes of lighting condition, view angle, and scale, researchers begun to explore applications of 3D information in human activity understanding in recently years. Unfortunately, features that work well on 2D videos usually don't perform well on 3D videos and there is no consensus on what 3D features should be used. Here we propose a model of human activity recognition based on 3D movements of body joints. Our method has three steps, learning dictionaries of sparse codes of 3D movements of joints, sparse coding, and classification. In the first step, space-time volumes of 3D movements of body joints are obtained via dense sampling and independent component analysis is then performed to construct a dictionary of sparse codes for each activity. In the second step, the space-time volumes are projected to the dictionaries and a set of sparse histograms of the projection coefficients are constructed as feature representations of the activities. Finally, the sparse histograms are used as inputs to a support vector machine to recognize human activities. We tested this model on three databases of human activities and found that it outperforms the state-of-the-art algorithms. Thus, this model can be used for real-time human activity recognition in many applications.
url http://europepmc.org/articles/PMC4256388?pdf=render
work_keys_str_mv AT jinqi learningdictionariesofsparsecodesof3dmovementsofbodyjointsforrealtimehumanactivityunderstanding
AT zhiyongyang learningdictionariesofsparsecodesof3dmovementsofbodyjointsforrealtimehumanactivityunderstanding
_version_ 1725103644277211136