Frequent Subspace Classifier

碩士 === 國立臺灣大學 === 資訊管理學研究所 === 97 === With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used...

Full description

Bibliographic Details
Main Authors: Ching-Wei Cheng, 鄭景瑋
Other Authors: 李瑞庭
Format: Others
Language:en_US
Published: 2009
Online Access:http://ndltd.ncl.edu.tw/handle/65886346622257718847
id ndltd-TW-097NTU05396036
record_format oai_dc
spelling ndltd-TW-097NTU053960362016-05-04T04:31:49Z http://ndltd.ncl.edu.tw/handle/65886346622257718847 Frequent Subspace Classifier 頻繁子空間之分類器 Ching-Wei Cheng 鄭景瑋 碩士 國立臺灣大學 資訊管理學研究所 97 With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used randomly-generated or some subspaces to construct a classifier. Therefore, in this thesis, we propose a hybrid classification method, called FSC (Frequent subspace classifier), to generate all potential subspaces and utilize these subspaces to construct a classifier. Our proposed method consists of three phases. First, we apply the discrete wavelet transform to reduce the dimensions of feature vectors. Next, we employ the frequent subspaces mining method to derive all potential subspaces. Finally, we exploit AdaBoost to select the significant subspaces from the potential subspaces derived to construct an ensemble classifier. Since the FSC generates all potential subspaces and selects the subspaces based on the maximum entropy reduction, it provides more opportunities to construct an effective classifier. The experiment results show that the FSC outperforms the SVM and LogitBoost in both UCI and stock datasets. 李瑞庭 2009 學位論文 ; thesis 32 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立臺灣大學 === 資訊管理學研究所 === 97 === With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used randomly-generated or some subspaces to construct a classifier. Therefore, in this thesis, we propose a hybrid classification method, called FSC (Frequent subspace classifier), to generate all potential subspaces and utilize these subspaces to construct a classifier. Our proposed method consists of three phases. First, we apply the discrete wavelet transform to reduce the dimensions of feature vectors. Next, we employ the frequent subspaces mining method to derive all potential subspaces. Finally, we exploit AdaBoost to select the significant subspaces from the potential subspaces derived to construct an ensemble classifier. Since the FSC generates all potential subspaces and selects the subspaces based on the maximum entropy reduction, it provides more opportunities to construct an effective classifier. The experiment results show that the FSC outperforms the SVM and LogitBoost in both UCI and stock datasets.
author2 李瑞庭
author_facet 李瑞庭
Ching-Wei Cheng
鄭景瑋
author Ching-Wei Cheng
鄭景瑋
spellingShingle Ching-Wei Cheng
鄭景瑋
Frequent Subspace Classifier
author_sort Ching-Wei Cheng
title Frequent Subspace Classifier
title_short Frequent Subspace Classifier
title_full Frequent Subspace Classifier
title_fullStr Frequent Subspace Classifier
title_full_unstemmed Frequent Subspace Classifier
title_sort frequent subspace classifier
publishDate 2009
url http://ndltd.ncl.edu.tw/handle/65886346622257718847
work_keys_str_mv AT chingweicheng frequentsubspaceclassifier
AT zhèngjǐngwěi frequentsubspaceclassifier
AT chingweicheng pínfánzikōngjiānzhīfēnlèiqì
AT zhèngjǐngwěi pínfánzikōngjiānzhīfēnlèiqì
_version_ 1718259931784675328