DeepPear: Deep Pose Estimation and Action Recognition
碩士 === 國立交通大學 === 多媒體工程研究所 === 107 === Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is dif...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2019
|
Online Access: | http://ndltd.ncl.edu.tw/handle/52d4yz |
id |
ndltd-TW-107NCTU5641044 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-107NCTU56410442019-11-26T05:16:54Z http://ndltd.ncl.edu.tw/handle/52d4yz DeepPear: Deep Pose Estimation and Action Recognition 人體骨架預測與基於骨架之動作識別 Jhuang, You-Ying 莊侑穎 碩士 國立交通大學 多媒體工程研究所 107 Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is difficult because the learning of actions could be affected by the cluttered background. In contrast to most video-based action recognition approaches which use 3D convolutional neural networks, the proposed method estimates 3D human pose first which can help remove the cluttered background and focus on human body. This makes the action learning not be overfitted by the cluttered background. Besides human pose, the proposed method also utilizes the RGB features nearby the predicted human joints to make our action prediction context-aware. After human pose estimation and RGB feature extraction, the proposed method uses a two-stream architecture to handle action recognition. Experimental results show that the proposed method outperformed many state-of-the-arts on NTU RGB+D which is a large-scale human action recognition dataset. Tsai, Wen-Jiin Chen, Hua-Tsung 蔡文錦 陳華總 2019 學位論文 ; thesis 31 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立交通大學 === 多媒體工程研究所 === 107 === Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is difficult because the learning of actions could be affected by the cluttered background. In contrast to most video-based action recognition approaches which use 3D convolutional neural networks, the proposed method estimates 3D human pose first which can help remove the cluttered background and focus on human body. This makes the action learning not be overfitted by the cluttered background. Besides human pose, the proposed method also utilizes the RGB features nearby the predicted human joints to make our action prediction context-aware. After human pose estimation and RGB feature extraction, the proposed method uses a two-stream architecture to handle action recognition. Experimental results show that the proposed method outperformed many state-of-the-arts on NTU RGB+D which is a large-scale human action recognition dataset.
|
author2 |
Tsai, Wen-Jiin |
author_facet |
Tsai, Wen-Jiin Jhuang, You-Ying 莊侑穎 |
author |
Jhuang, You-Ying 莊侑穎 |
spellingShingle |
Jhuang, You-Ying 莊侑穎 DeepPear: Deep Pose Estimation and Action Recognition |
author_sort |
Jhuang, You-Ying |
title |
DeepPear: Deep Pose Estimation and Action Recognition |
title_short |
DeepPear: Deep Pose Estimation and Action Recognition |
title_full |
DeepPear: Deep Pose Estimation and Action Recognition |
title_fullStr |
DeepPear: Deep Pose Estimation and Action Recognition |
title_full_unstemmed |
DeepPear: Deep Pose Estimation and Action Recognition |
title_sort |
deeppear: deep pose estimation and action recognition |
publishDate |
2019 |
url |
http://ndltd.ncl.edu.tw/handle/52d4yz |
work_keys_str_mv |
AT jhuangyouying deeppeardeepposeestimationandactionrecognition AT zhuāngyòuyǐng deeppeardeepposeestimationandactionrecognition AT jhuangyouying réntǐgǔjiàyùcèyǔjīyúgǔjiàzhīdòngzuòshíbié AT zhuāngyòuyǐng réntǐgǔjiàyùcèyǔjīyúgǔjiàzhīdòngzuòshíbié |
_version_ |
1719296548135763968 |