Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes
Supervised learning is a machine learning technique used for creating a data prediction model. This article focuses on finding high performance supervised learning algorithms with varied training data sizes, varied number of attributes, and time spent on prediction. This studied evaluated seven algo...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Khon Kaen University
2018-09-01
|
Series: | Engineering and Applied Science Research |
Subjects: | |
Online Access: | https://www.tci-thaijo.org/index.php/easr/article/download/88019/107554/ |
id |
doaj-75918f5ecb2642959d406e59051d1785 |
---|---|
record_format |
Article |
spelling |
doaj-75918f5ecb2642959d406e59051d17852020-11-24T22:19:46ZengKhon Kaen UniversityEngineering and Applied Science Research2539-61612539-62182018-09-0145322122910.14456/easr.2018.28Performance evaluation of supervised learning algorithms with various training data sizes and missing attributesChaluemwut NoyunsanTatpong KatanyukulKanda SaikaewSupervised learning is a machine learning technique used for creating a data prediction model. This article focuses on finding high performance supervised learning algorithms with varied training data sizes, varied number of attributes, and time spent on prediction. This studied evaluated seven algorithms, Boosting, Random Forest, Bagging, Naive Bayes, K-Nearest Neighbours (K-NN), Decision Tree, and Support Vector Machine (SVM), on seven data sets that are the standard benchmark from University of California, Irvine (UCI) with two evaluation metrics and experimental settings of various training data sizes and missing key attributes. Our findings reveal that Bagging, Random Forest, and SVM are overall the three most accurate algorithms. However, when presence of key attribute values is of concern, K-NN is recommended as its performance is affected the least. Alternatively, when training data sizes may be not large enough, Naive Bayes is preferable since it is the most insensitive algorithm to training data sizes. The algorithms are characterized on a two-dimension chart based on prediction performance and computation time. This chart is expected to guide a novice user to choose an appropriate method for his/her demand. Based on this chart, in general, Bagging and Random Forest are the two most recommended algorithms because of their high performance and speed.https://www.tci-thaijo.org/index.php/easr/article/download/88019/107554/Supervised learning algorithmsEvaluation metricsPerformance comparison |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Chaluemwut Noyunsan Tatpong Katanyukul Kanda Saikaew |
spellingShingle |
Chaluemwut Noyunsan Tatpong Katanyukul Kanda Saikaew Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes Engineering and Applied Science Research Supervised learning algorithms Evaluation metrics Performance comparison |
author_facet |
Chaluemwut Noyunsan Tatpong Katanyukul Kanda Saikaew |
author_sort |
Chaluemwut Noyunsan |
title |
Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
title_short |
Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
title_full |
Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
title_fullStr |
Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
title_full_unstemmed |
Performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
title_sort |
performance evaluation of supervised learning algorithms with various training data sizes and missing attributes |
publisher |
Khon Kaen University |
series |
Engineering and Applied Science Research |
issn |
2539-6161 2539-6218 |
publishDate |
2018-09-01 |
description |
Supervised learning is a machine learning technique used for creating a data prediction model. This article focuses on finding high performance supervised learning algorithms with varied training data sizes, varied number of attributes, and time spent on prediction. This studied evaluated seven algorithms, Boosting, Random Forest, Bagging, Naive Bayes, K-Nearest Neighbours (K-NN), Decision Tree, and Support Vector Machine (SVM), on seven data sets that are the standard benchmark from University of California, Irvine (UCI) with two evaluation metrics and experimental settings of various training data sizes and missing key attributes. Our findings reveal that Bagging, Random Forest, and SVM are overall the three most accurate algorithms. However, when presence of key attribute values is of concern, K-NN is recommended as its performance is affected the least. Alternatively, when training data sizes may be not large enough, Naive Bayes is preferable since it is the most insensitive algorithm to training data sizes. The algorithms are characterized on a two-dimension chart based on prediction performance and computation time. This chart is expected to guide a novice user to choose an appropriate method for his/her demand. Based on this chart, in general, Bagging and Random Forest are the two most recommended algorithms because of their high performance and speed. |
topic |
Supervised learning algorithms Evaluation metrics Performance comparison |
url |
https://www.tci-thaijo.org/index.php/easr/article/download/88019/107554/ |
work_keys_str_mv |
AT chaluemwutnoyunsan performanceevaluationofsupervisedlearningalgorithmswithvarioustrainingdatasizesandmissingattributes AT tatpongkatanyukul performanceevaluationofsupervisedlearningalgorithmswithvarioustrainingdatasizesandmissingattributes AT kandasaikaew performanceevaluationofsupervisedlearningalgorithmswithvarioustrainingdatasizesandmissingattributes |
_version_ |
1725777424914841600 |