The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
碩士 === 華梵大學 === 資訊管理學系碩士班 === 96 === Traditionally, decision tree performs as not well as novel classifiers like BPN or SVM in aspect of accuracy. However, the human readable result generated by decision tree truly makes it capable to fulfill the requirement of some expert domain like medical、busine...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2008
|
Online Access: | http://ndltd.ncl.edu.tw/handle/98803818878451365043 |
id |
ndltd-TW-096HCHT0396054 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-096HCHT03960542015-10-13T13:47:50Z http://ndltd.ncl.edu.tw/handle/98803818878451365043 The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree 利用模擬退火法進行決策樹中連續型屬性分割點最佳化之研究 Ke Chih Ning 甯格致 碩士 華梵大學 資訊管理學系碩士班 96 Traditionally, decision tree performs as not well as novel classifiers like BPN or SVM in aspect of accuracy. However, the human readable result generated by decision tree truly makes it capable to fulfill the requirement of some expert domain like medical、business…etc. But when the dataset has severe data confliction in it or there are large amount of information hidden in it, the decision tree will turns out a big size tree with high depth which makes it hard to understand by human. Although decision tree yields out average performance among usual applications, the inefficiency of binary cut which used to split the continuous data is still an academic topic, and there are several methods claims to facilitate the performance of continuous attribute splitting, but eventually the improvement lacks of generalization. In this research, we proposed a meta-heuristics method (e.g. simulated annealing) with a proper designed Objective Function and two pruning method (In-Build Pruning, Post Pruning) in the process the tree generation were joined. We choose 10 UCI datasets as our original data. All of the experiments were follow the procedure of 10-folds cross validation method. Our proposed method finally generated much more simplified decision trees with relatively small tree size & almost half of the tree depth compared with C4.5. In addition, the accuracy has no significant differences. T.Y. Tseng 曾綜源 2008 學位論文 ; thesis 77 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 華梵大學 === 資訊管理學系碩士班 === 96 === Traditionally, decision tree performs as not well as novel classifiers like BPN or
SVM in aspect of accuracy. However, the human readable result generated by
decision tree truly makes it capable to fulfill the requirement of some expert domain
like medical、business…etc. But when the dataset has severe data confliction in it or
there are large amount of information hidden in it, the decision tree will turns out a
big size tree with high depth which makes it hard to understand by human. Although
decision tree yields out average performance among usual applications, the
inefficiency of binary cut which used to split the continuous data is still an academic
topic, and there are several methods claims to facilitate the performance of continuous
attribute splitting, but eventually the improvement lacks of generalization. In this
research, we proposed a meta-heuristics method (e.g. simulated annealing) with a
proper designed Objective Function and two pruning method (In-Build Pruning, Post
Pruning) in the process the tree generation were joined. We choose 10 UCI datasets as
our original data. All of the experiments were follow the procedure of 10-folds cross
validation method. Our proposed method finally generated much more simplified
decision trees with relatively small tree size & almost half of the tree depth compared
with C4.5. In addition, the accuracy has no significant differences.
|
author2 |
T.Y. Tseng |
author_facet |
T.Y. Tseng Ke Chih Ning 甯格致 |
author |
Ke Chih Ning 甯格致 |
spellingShingle |
Ke Chih Ning 甯格致 The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
author_sort |
Ke Chih Ning |
title |
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
title_short |
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
title_full |
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
title_fullStr |
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
title_full_unstemmed |
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree |
title_sort |
studies of optimal continuous attribute multiple splitting in decision tree |
publishDate |
2008 |
url |
http://ndltd.ncl.edu.tw/handle/98803818878451365043 |
work_keys_str_mv |
AT kechihning thestudiesofoptimalcontinuousattributemultiplesplittingindecisiontree AT nínggézhì thestudiesofoptimalcontinuousattributemultiplesplittingindecisiontree AT kechihning lìyòngmónǐtuìhuǒfǎjìnxíngjuécèshùzhōngliánxùxíngshǔxìngfēngēdiǎnzuìjiāhuàzhīyánjiū AT nínggézhì lìyòngmónǐtuìhuǒfǎjìnxíngjuécèshùzhōngliánxùxíngshǔxìngfēngēdiǎnzuìjiāhuàzhīyánjiū AT kechihning studiesofoptimalcontinuousattributemultiplesplittingindecisiontree AT nínggézhì studiesofoptimalcontinuousattributemultiplesplittingindecisiontree |
_version_ |
1717742435153477632 |