Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints
碩士 === 國立臺灣大學 === 資訊工程學研究所 === 103 === There are a few existing annotation systems that aim to provide a platform for video annotation. Most of them focus on activity annotation while others concentrate on labeling individual objects. However, the latters focus on only labeling objects with bounding...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2014
|
Online Access: | http://ndltd.ncl.edu.tw/handle/68845600289225045103 |
id |
ndltd-TW-103NTU05392007 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-103NTU053920072016-07-02T04:21:18Z http://ndltd.ncl.edu.tw/handle/68845600289225045103 Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints 以非剛性對齊演算法與使用者標記之明顯特徵對應關係建立影片資料之稠密對應 Yen-Ting Chen 陳彥廷 碩士 國立臺灣大學 資訊工程學研究所 103 There are a few existing annotation systems that aim to provide a platform for video annotation. Most of them focus on activity annotation while others concentrate on labeling individual objects. However, the latters focus on only labeling objects with bounding boxes or only using interpolation techniques to help user labeling. Moreover, only one of them try to find the dense correspondence inside the object contour. Issues of dense correspondences annotation across video frames are not well addressed yet. Inspired by this, a video annotation system that focuses on dense correspondences annotation inside the object contour is proposed in this work. In addition, since labeling detail object contour and dense correspondences across a whole video is a daunting task, we also minimize user''s effort by applying an interactive segmentation and tracking algorithm that utilizes information from optical flow and edges that helps the user easier to observe the salient feature correspondences between two video frames. Edges could help the user to find out the detail contour or local patterns of the object. The user is required to check and modify the salient feature correspondences obtained by the algorithm. Dense correspondences in the textureless region are extracted by a non-rigid registration algorithm from the salient feature correspondences verified by the user. The user only needs to label the first frame of the video and correct some minor errors in the subsequent frames for the whole video annotation. The result shows that the proposed framework is more suitable to label non-rigid objects. Chieh-Chih Wang 王傑智 2014 學位論文 ; thesis 36 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺灣大學 === 資訊工程學研究所 === 103 === There are a few existing annotation systems that aim to provide a platform for video annotation. Most of them focus on activity annotation while others concentrate on labeling individual objects. However, the latters focus on only labeling objects with bounding boxes or only using interpolation techniques to help user labeling. Moreover, only one of them try to find the dense correspondence inside the object contour. Issues of dense correspondences annotation across video frames are not well addressed yet. Inspired by this, a video annotation system that focuses on dense correspondences annotation inside the object contour is proposed in this work. In addition, since labeling detail object contour and dense correspondences across a whole video is a daunting task, we also minimize user''s effort by applying an interactive segmentation and tracking algorithm that utilizes information from optical flow and edges that helps the user easier to observe the salient feature correspondences between two video frames. Edges could help the user to find out the detail contour or local patterns of the object. The user is required to check and modify the salient feature correspondences obtained by the algorithm. Dense correspondences in the textureless region are extracted by a non-rigid registration algorithm from the salient feature correspondences verified by the user. The user only needs to label the first frame of the video and correct some minor errors in the subsequent frames for the whole video annotation. The result shows that the proposed framework is more suitable to label non-rigid objects.
|
author2 |
Chieh-Chih Wang |
author_facet |
Chieh-Chih Wang Yen-Ting Chen 陳彥廷 |
author |
Yen-Ting Chen 陳彥廷 |
spellingShingle |
Yen-Ting Chen 陳彥廷 Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
author_sort |
Yen-Ting Chen |
title |
Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
title_short |
Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
title_full |
Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
title_fullStr |
Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
title_full_unstemmed |
Dense Correspondence Annotation of Video Data Using Non-Rigid Registration with Salient Feature Correspondence Constraints |
title_sort |
dense correspondence annotation of video data using non-rigid registration with salient feature correspondence constraints |
publishDate |
2014 |
url |
http://ndltd.ncl.edu.tw/handle/68845600289225045103 |
work_keys_str_mv |
AT yentingchen densecorrespondenceannotationofvideodatausingnonrigidregistrationwithsalientfeaturecorrespondenceconstraints AT chényàntíng densecorrespondenceannotationofvideodatausingnonrigidregistrationwithsalientfeaturecorrespondenceconstraints AT yentingchen yǐfēigāngxìngduìqíyǎnsuànfǎyǔshǐyòngzhěbiāojìzhīmíngxiǎntèzhēngduìyīngguānxìjiànlìyǐngpiànzīliàozhīchóumìduìyīng AT chényàntíng yǐfēigāngxìngduìqíyǎnsuànfǎyǔshǐyòngzhěbiāojìzhīmíngxiǎntèzhēngduìyīngguānxìjiànlìyǐngpiànzīliàozhīchóumìduìyīng |
_version_ |
1718332788507148288 |