The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree

碩士 === 華梵大學 === 資訊管理學系碩士班 === 96 === Traditionally, decision tree performs as not well as novel classifiers like BPN or SVM in aspect of accuracy. However, the human readable result generated by decision tree truly makes it capable to fulfill the requirement of some expert domain like medical、busine...

Full description

Bibliographic Details
Main Authors: Ke Chih Ning, 甯格致
Other Authors: T.Y. Tseng
Format: Others
Language:zh-TW
Published: 2008
Online Access:http://ndltd.ncl.edu.tw/handle/98803818878451365043
id ndltd-TW-096HCHT0396054
record_format oai_dc
spelling ndltd-TW-096HCHT03960542015-10-13T13:47:50Z http://ndltd.ncl.edu.tw/handle/98803818878451365043 The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree 利用模擬退火法進行決策樹中連續型屬性分割點最佳化之研究 Ke Chih Ning 甯格致 碩士 華梵大學 資訊管理學系碩士班 96 Traditionally, decision tree performs as not well as novel classifiers like BPN or SVM in aspect of accuracy. However, the human readable result generated by decision tree truly makes it capable to fulfill the requirement of some expert domain like medical、business…etc. But when the dataset has severe data confliction in it or there are large amount of information hidden in it, the decision tree will turns out a big size tree with high depth which makes it hard to understand by human. Although decision tree yields out average performance among usual applications, the inefficiency of binary cut which used to split the continuous data is still an academic topic, and there are several methods claims to facilitate the performance of continuous attribute splitting, but eventually the improvement lacks of generalization. In this research, we proposed a meta-heuristics method (e.g. simulated annealing) with a proper designed Objective Function and two pruning method (In-Build Pruning, Post Pruning) in the process the tree generation were joined. We choose 10 UCI datasets as our original data. All of the experiments were follow the procedure of 10-folds cross validation method. Our proposed method finally generated much more simplified decision trees with relatively small tree size & almost half of the tree depth compared with C4.5. In addition, the accuracy has no significant differences. T.Y. Tseng 曾綜源 2008 學位論文 ; thesis 77 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 華梵大學 === 資訊管理學系碩士班 === 96 === Traditionally, decision tree performs as not well as novel classifiers like BPN or SVM in aspect of accuracy. However, the human readable result generated by decision tree truly makes it capable to fulfill the requirement of some expert domain like medical、business…etc. But when the dataset has severe data confliction in it or there are large amount of information hidden in it, the decision tree will turns out a big size tree with high depth which makes it hard to understand by human. Although decision tree yields out average performance among usual applications, the inefficiency of binary cut which used to split the continuous data is still an academic topic, and there are several methods claims to facilitate the performance of continuous attribute splitting, but eventually the improvement lacks of generalization. In this research, we proposed a meta-heuristics method (e.g. simulated annealing) with a proper designed Objective Function and two pruning method (In-Build Pruning, Post Pruning) in the process the tree generation were joined. We choose 10 UCI datasets as our original data. All of the experiments were follow the procedure of 10-folds cross validation method. Our proposed method finally generated much more simplified decision trees with relatively small tree size & almost half of the tree depth compared with C4.5. In addition, the accuracy has no significant differences.
author2 T.Y. Tseng
author_facet T.Y. Tseng
Ke Chih Ning
甯格致
author Ke Chih Ning
甯格致
spellingShingle Ke Chih Ning
甯格致
The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
author_sort Ke Chih Ning
title The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
title_short The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
title_full The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
title_fullStr The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
title_full_unstemmed The Studies of Optimal Continuous Attribute Multiple Splitting in Decision Tree
title_sort studies of optimal continuous attribute multiple splitting in decision tree
publishDate 2008
url http://ndltd.ncl.edu.tw/handle/98803818878451365043
work_keys_str_mv AT kechihning thestudiesofoptimalcontinuousattributemultiplesplittingindecisiontree
AT nínggézhì thestudiesofoptimalcontinuousattributemultiplesplittingindecisiontree
AT kechihning lìyòngmónǐtuìhuǒfǎjìnxíngjuécèshùzhōngliánxùxíngshǔxìngfēngēdiǎnzuìjiāhuàzhīyánjiū
AT nínggézhì lìyòngmónǐtuìhuǒfǎjìnxíngjuécèshùzhōngliánxùxíngshǔxìngfēngēdiǎnzuìjiāhuàzhīyánjiū
AT kechihning studiesofoptimalcontinuousattributemultiplesplittingindecisiontree
AT nínggézhì studiesofoptimalcontinuousattributemultiplesplittingindecisiontree
_version_ 1717742435153477632