Weighted Scoring in Geometric Space for Decision Tree Ensemble
In order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training data...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9079594/ |
id |
doaj-26d0ba758c9a48c181568a7c3e51f3de |
---|---|
record_format |
Article |
spelling |
doaj-26d0ba758c9a48c181568a7c3e51f3de2021-03-30T01:44:22ZengIEEEIEEE Access2169-35362020-01-018821008210710.1109/ACCESS.2020.29907219079594Weighted Scoring in Geometric Space for Decision Tree EnsembleJedrzej Biedrzycki0https://orcid.org/0000-0002-4924-1759Robert Burduk1https://orcid.org/0000-0002-3506-6611Faculty of Electronics, Wroclaw University of Science and Technology, Wroclaw, PolandFaculty of Electronics, Wroclaw University of Science and Technology, Wroclaw, PolandIn order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training dataset. In this paper, we propose the algorithm that uses horizontal partitioning the learning set and uses decision trees as base models to obtain decision regions. In the proposed approach feature space is divided into disjoint subspace. Additionally, the location of the subspace centroids, as well as the size and location of decision regions, are used in order to determine the weights needed in the last process of creating MCS, i.e. in the integration phase. The proposed algorithm was evaluated employing multiple open-source benchmarking datasets, compared using accuracy and Matthews correlation coefficient performance measures with two existing MCS methods - random forest and majority voting. The statistical analysis confirms an improvement in recognition compared to the random forest. In addition, we proved that for infinitely dense space division proposed algorithm is equivalent to majority voting.https://ieeexplore.ieee.org/document/9079594/Decision treeensemble classifiermajority votingmultiple classifier systemrandom forest |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Jedrzej Biedrzycki Robert Burduk |
spellingShingle |
Jedrzej Biedrzycki Robert Burduk Weighted Scoring in Geometric Space for Decision Tree Ensemble IEEE Access Decision tree ensemble classifier majority voting multiple classifier system random forest |
author_facet |
Jedrzej Biedrzycki Robert Burduk |
author_sort |
Jedrzej Biedrzycki |
title |
Weighted Scoring in Geometric Space for Decision Tree Ensemble |
title_short |
Weighted Scoring in Geometric Space for Decision Tree Ensemble |
title_full |
Weighted Scoring in Geometric Space for Decision Tree Ensemble |
title_fullStr |
Weighted Scoring in Geometric Space for Decision Tree Ensemble |
title_full_unstemmed |
Weighted Scoring in Geometric Space for Decision Tree Ensemble |
title_sort |
weighted scoring in geometric space for decision tree ensemble |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
In order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training dataset. In this paper, we propose the algorithm that uses horizontal partitioning the learning set and uses decision trees as base models to obtain decision regions. In the proposed approach feature space is divided into disjoint subspace. Additionally, the location of the subspace centroids, as well as the size and location of decision regions, are used in order to determine the weights needed in the last process of creating MCS, i.e. in the integration phase. The proposed algorithm was evaluated employing multiple open-source benchmarking datasets, compared using accuracy and Matthews correlation coefficient performance measures with two existing MCS methods - random forest and majority voting. The statistical analysis confirms an improvement in recognition compared to the random forest. In addition, we proved that for infinitely dense space division proposed algorithm is equivalent to majority voting. |
topic |
Decision tree ensemble classifier majority voting multiple classifier system random forest |
url |
https://ieeexplore.ieee.org/document/9079594/ |
work_keys_str_mv |
AT jedrzejbiedrzycki weightedscoringingeometricspacefordecisiontreeensemble AT robertburduk weightedscoringingeometricspacefordecisiontreeensemble |
_version_ |
1724186412230639616 |