Frequent Subspace Classifier
碩士 === 國立臺灣大學 === 資訊管理學研究所 === 97 === With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2009
|
Online Access: | http://ndltd.ncl.edu.tw/handle/65886346622257718847 |
id |
ndltd-TW-097NTU05396036 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-097NTU053960362016-05-04T04:31:49Z http://ndltd.ncl.edu.tw/handle/65886346622257718847 Frequent Subspace Classifier 頻繁子空間之分類器 Ching-Wei Cheng 鄭景瑋 碩士 國立臺灣大學 資訊管理學研究所 97 With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used randomly-generated or some subspaces to construct a classifier. Therefore, in this thesis, we propose a hybrid classification method, called FSC (Frequent subspace classifier), to generate all potential subspaces and utilize these subspaces to construct a classifier. Our proposed method consists of three phases. First, we apply the discrete wavelet transform to reduce the dimensions of feature vectors. Next, we employ the frequent subspaces mining method to derive all potential subspaces. Finally, we exploit AdaBoost to select the significant subspaces from the potential subspaces derived to construct an ensemble classifier. Since the FSC generates all potential subspaces and selects the subspaces based on the maximum entropy reduction, it provides more opportunities to construct an effective classifier. The experiment results show that the FSC outperforms the SVM and LogitBoost in both UCI and stock datasets. 李瑞庭 2009 學位論文 ; thesis 32 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺灣大學 === 資訊管理學研究所 === 97 === With the amount of the data increasing rapidly, it is infeasible to consider all the dimensions of the data to perform classification. Thus, constructing a classifier based on subspaces has attracted more and more attention. The previously proposed methods used randomly-generated or some subspaces to construct a classifier. Therefore, in this thesis, we propose a hybrid classification method, called FSC (Frequent subspace classifier), to generate all potential subspaces and utilize these subspaces to construct a classifier. Our proposed method consists of three phases. First, we apply the discrete wavelet transform to reduce the dimensions of feature vectors. Next, we employ the frequent subspaces mining method to derive all potential subspaces. Finally, we exploit AdaBoost to select the significant subspaces from the potential subspaces derived to construct an ensemble classifier. Since the FSC generates all potential subspaces and selects the subspaces based on the maximum entropy reduction, it provides more opportunities to construct an effective classifier. The experiment results show that the FSC outperforms the SVM and LogitBoost in both UCI and stock datasets.
|
author2 |
李瑞庭 |
author_facet |
李瑞庭 Ching-Wei Cheng 鄭景瑋 |
author |
Ching-Wei Cheng 鄭景瑋 |
spellingShingle |
Ching-Wei Cheng 鄭景瑋 Frequent Subspace Classifier |
author_sort |
Ching-Wei Cheng |
title |
Frequent Subspace Classifier |
title_short |
Frequent Subspace Classifier |
title_full |
Frequent Subspace Classifier |
title_fullStr |
Frequent Subspace Classifier |
title_full_unstemmed |
Frequent Subspace Classifier |
title_sort |
frequent subspace classifier |
publishDate |
2009 |
url |
http://ndltd.ncl.edu.tw/handle/65886346622257718847 |
work_keys_str_mv |
AT chingweicheng frequentsubspaceclassifier AT zhèngjǐngwěi frequentsubspaceclassifier AT chingweicheng pínfánzikōngjiānzhīfēnlèiqì AT zhèngjǐngwěi pínfánzikōngjiānzhīfēnlèiqì |
_version_ |
1718259931784675328 |