Human action recognition based on HOIRM feature fusion and AP clustering BOW.
In this paper, we propose a human action recognition method using HOIRM (histogram of oriented interest region motion) feature fusion and a BOW (bag of words) model based on AP (affinity propagation) clustering. First, a HOIRM feature extraction method based on spatiotemporal interest points ROI is...
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2019-01-01
|
Series: | PLoS ONE |
Online Access: | https://doi.org/10.1371/journal.pone.0219910 |
id |
doaj-9c0463c665364887b366cee08ac9ecdd |
---|---|
record_format |
Article |
spelling |
doaj-9c0463c665364887b366cee08ac9ecdd2021-03-03T20:33:25ZengPublic Library of Science (PLoS)PLoS ONE1932-62032019-01-01147e021991010.1371/journal.pone.0219910Human action recognition based on HOIRM feature fusion and AP clustering BOW.Ruo-Hong HuanChao-Jie XieFeng GuoKai-Kai ChiKe-Ji MaoYing-Long LiYun PanIn this paper, we propose a human action recognition method using HOIRM (histogram of oriented interest region motion) feature fusion and a BOW (bag of words) model based on AP (affinity propagation) clustering. First, a HOIRM feature extraction method based on spatiotemporal interest points ROI is proposed. HOIRM can be regarded as a middle-level feature between local and global features. Then, HOIRM is fused with 3D HOG and 3D HOF local features using a cumulative histogram. The method further improves the robustness of local features to camera view angle and distance variations in complex scenes, which in turn improves the correct rate of action recognition. Finally, a BOW model based on AP clustering is proposed and applied to action classification. It obtains the appropriate visual dictionary capacity and achieves better clustering effect for the joint description of a variety of features. The experimental results demonstrate that by using the fused features with the proposed BOW model, the average recognition rate is 95.75% in the KTH database, and 88.25% in the UCF database, which are both higher than those by using only 3D HOG+3D HOF or HOIRM features. Moreover, the average recognition rate achieved by the proposed method in the two databases is higher than that obtained by other methods.https://doi.org/10.1371/journal.pone.0219910 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Ruo-Hong Huan Chao-Jie Xie Feng Guo Kai-Kai Chi Ke-Ji Mao Ying-Long Li Yun Pan |
spellingShingle |
Ruo-Hong Huan Chao-Jie Xie Feng Guo Kai-Kai Chi Ke-Ji Mao Ying-Long Li Yun Pan Human action recognition based on HOIRM feature fusion and AP clustering BOW. PLoS ONE |
author_facet |
Ruo-Hong Huan Chao-Jie Xie Feng Guo Kai-Kai Chi Ke-Ji Mao Ying-Long Li Yun Pan |
author_sort |
Ruo-Hong Huan |
title |
Human action recognition based on HOIRM feature fusion and AP clustering BOW. |
title_short |
Human action recognition based on HOIRM feature fusion and AP clustering BOW. |
title_full |
Human action recognition based on HOIRM feature fusion and AP clustering BOW. |
title_fullStr |
Human action recognition based on HOIRM feature fusion and AP clustering BOW. |
title_full_unstemmed |
Human action recognition based on HOIRM feature fusion and AP clustering BOW. |
title_sort |
human action recognition based on hoirm feature fusion and ap clustering bow. |
publisher |
Public Library of Science (PLoS) |
series |
PLoS ONE |
issn |
1932-6203 |
publishDate |
2019-01-01 |
description |
In this paper, we propose a human action recognition method using HOIRM (histogram of oriented interest region motion) feature fusion and a BOW (bag of words) model based on AP (affinity propagation) clustering. First, a HOIRM feature extraction method based on spatiotemporal interest points ROI is proposed. HOIRM can be regarded as a middle-level feature between local and global features. Then, HOIRM is fused with 3D HOG and 3D HOF local features using a cumulative histogram. The method further improves the robustness of local features to camera view angle and distance variations in complex scenes, which in turn improves the correct rate of action recognition. Finally, a BOW model based on AP clustering is proposed and applied to action classification. It obtains the appropriate visual dictionary capacity and achieves better clustering effect for the joint description of a variety of features. The experimental results demonstrate that by using the fused features with the proposed BOW model, the average recognition rate is 95.75% in the KTH database, and 88.25% in the UCF database, which are both higher than those by using only 3D HOG+3D HOF or HOIRM features. Moreover, the average recognition rate achieved by the proposed method in the two databases is higher than that obtained by other methods. |
url |
https://doi.org/10.1371/journal.pone.0219910 |
work_keys_str_mv |
AT ruohonghuan humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT chaojiexie humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT fengguo humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT kaikaichi humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT kejimao humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT yinglongli humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow AT yunpan humanactionrecognitionbasedonhoirmfeaturefusionandapclusteringbow |
_version_ |
1714821906774884352 |