DeepPear: Deep Pose Estimation and Action Recognition

碩士 === 國立交通大學 === 多媒體工程研究所 === 107 === Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is dif...

Full description

Bibliographic Details
Main Authors: Jhuang, You-Ying, 莊侑穎
Other Authors: Tsai, Wen-Jiin
Format: Others
Language:en_US
Published: 2019
Online Access:http://ndltd.ncl.edu.tw/handle/52d4yz
id ndltd-TW-107NCTU5641044
record_format oai_dc
spelling ndltd-TW-107NCTU56410442019-11-26T05:16:54Z http://ndltd.ncl.edu.tw/handle/52d4yz DeepPear: Deep Pose Estimation and Action Recognition 人體骨架預測與基於骨架之動作識別 Jhuang, You-Ying 莊侑穎 碩士 國立交通大學 多媒體工程研究所 107 Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is difficult because the learning of actions could be affected by the cluttered background. In contrast to most video-based action recognition approaches which use 3D convolutional neural networks, the proposed method estimates 3D human pose first which can help remove the cluttered background and focus on human body. This makes the action learning not be overfitted by the cluttered background. Besides human pose, the proposed method also utilizes the RGB features nearby the predicted human joints to make our action prediction context-aware. After human pose estimation and RGB feature extraction, the proposed method uses a two-stream architecture to handle action recognition. Experimental results show that the proposed method outperformed many state-of-the-arts on NTU RGB+D which is a large-scale human action recognition dataset. Tsai, Wen-Jiin Chen, Hua-Tsung 蔡文錦 陳華總 2019 學位論文 ; thesis 31 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立交通大學 === 多媒體工程研究所 === 107 === Over the last few years, human action recognition has been a popular issue since this task can be applied in multiple applications such as intelligent surveillance systems, autonomous vehicle control and robotics. Human action recognition using RGB video is difficult because the learning of actions could be affected by the cluttered background. In contrast to most video-based action recognition approaches which use 3D convolutional neural networks, the proposed method estimates 3D human pose first which can help remove the cluttered background and focus on human body. This makes the action learning not be overfitted by the cluttered background. Besides human pose, the proposed method also utilizes the RGB features nearby the predicted human joints to make our action prediction context-aware. After human pose estimation and RGB feature extraction, the proposed method uses a two-stream architecture to handle action recognition. Experimental results show that the proposed method outperformed many state-of-the-arts on NTU RGB+D which is a large-scale human action recognition dataset.
author2 Tsai, Wen-Jiin
author_facet Tsai, Wen-Jiin
Jhuang, You-Ying
莊侑穎
author Jhuang, You-Ying
莊侑穎
spellingShingle Jhuang, You-Ying
莊侑穎
DeepPear: Deep Pose Estimation and Action Recognition
author_sort Jhuang, You-Ying
title DeepPear: Deep Pose Estimation and Action Recognition
title_short DeepPear: Deep Pose Estimation and Action Recognition
title_full DeepPear: Deep Pose Estimation and Action Recognition
title_fullStr DeepPear: Deep Pose Estimation and Action Recognition
title_full_unstemmed DeepPear: Deep Pose Estimation and Action Recognition
title_sort deeppear: deep pose estimation and action recognition
publishDate 2019
url http://ndltd.ncl.edu.tw/handle/52d4yz
work_keys_str_mv AT jhuangyouying deeppeardeepposeestimationandactionrecognition
AT zhuāngyòuyǐng deeppeardeepposeestimationandactionrecognition
AT jhuangyouying réntǐgǔjiàyùcèyǔjīyúgǔjiàzhīdòngzuòshíbié
AT zhuāngyòuyǐng réntǐgǔjiàyùcèyǔjīyúgǔjiàzhīdòngzuòshíbié
_version_ 1719296548135763968