Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data
The performance of many learning and data mining algorithms depends critically on suitable metrics to assess efficiency over the input space. Learning a suitable metric from examples may, therefore, be the key to successful application of these algorithms. We have demonstrated that the k-nearest nei...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
De Gruyter
2016-12-01
|
Series: | Journal of Integrative Bioinformatics |
Online Access: | https://doi.org/10.1515/jib-2016-304 |
id |
doaj-1cf35e1342104660af9a4c976ce6d912 |
---|---|
record_format |
Article |
spelling |
doaj-1cf35e1342104660af9a4c976ce6d9122021-09-06T19:40:32ZengDe GruyterJournal of Integrative Bioinformatics1613-45162016-12-01135112110.1515/jib-2016-304jib-2016-304Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs DataYousef Malik0Khalifa Waleed1AbdAllah Loai2Community Information Systems, Zefat Academic College, Zefat, 13206, IsraelComputer Science, The College of Sakhnin, Sakhnin, 30810, IsraelComputer Science, The College of Sakhnin, Sakhnin, 30810, IsraelThe performance of many learning and data mining algorithms depends critically on suitable metrics to assess efficiency over the input space. Learning a suitable metric from examples may, therefore, be the key to successful application of these algorithms. We have demonstrated that the k-nearest neighbor (kNN) classification can be significantly improved by learning a distance metric from labeled examples. The clustering ensemble is used to define the distance between points in respect to how they co-cluster. This distance is then used within the framework of the kNN algorithm to define a classifier named ensemble clustering kNN classifier (EC-kNN). In many instances in our experiments we achieved highest accuracy while SVM failed to perform as well. In this study, we compare the performance of a two-class classifier using EC-kNN with different one-class and two-class classifiers. The comparison was applied to seven different plant microRNA species considering eight feature selection methods. In this study, the averaged results show that EC-kNN outperforms all other methods employed here and previously published results for the same data. In conclusion, this study shows that the chosen classifier shows high performance when the distance metric is carefully chosen.https://doi.org/10.1515/jib-2016-304 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Yousef Malik Khalifa Waleed AbdAllah Loai |
spellingShingle |
Yousef Malik Khalifa Waleed AbdAllah Loai Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data Journal of Integrative Bioinformatics |
author_facet |
Yousef Malik Khalifa Waleed AbdAllah Loai |
author_sort |
Yousef Malik |
title |
Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data |
title_short |
Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data |
title_full |
Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data |
title_fullStr |
Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data |
title_full_unstemmed |
Ensemble Clustering Classification Applied to Competing SVM and One-Class Classifiers Exemplified by Plant MicroRNAs Data |
title_sort |
ensemble clustering classification applied to competing svm and one-class classifiers exemplified by plant micrornas data |
publisher |
De Gruyter |
series |
Journal of Integrative Bioinformatics |
issn |
1613-4516 |
publishDate |
2016-12-01 |
description |
The performance of many learning and data mining algorithms depends critically on suitable metrics to assess efficiency over the input space. Learning a suitable metric from examples may, therefore, be the key to successful application of these algorithms. We have demonstrated that the k-nearest neighbor (kNN) classification can be significantly improved by learning a distance metric from labeled examples. The clustering ensemble is used to define the distance between points in respect to how they co-cluster. This distance is then used within the framework of the kNN algorithm to define a classifier named ensemble clustering kNN classifier (EC-kNN). In many instances in our experiments we achieved highest accuracy while SVM failed to perform as well. In this study, we compare the performance of a two-class classifier using EC-kNN with different one-class and two-class classifiers. The comparison was applied to seven different plant microRNA species considering eight feature selection methods. In this study, the averaged results show that EC-kNN outperforms all other methods employed here and previously published results for the same data. In conclusion, this study shows that the chosen classifier shows high performance when the distance metric is carefully chosen. |
url |
https://doi.org/10.1515/jib-2016-304 |
work_keys_str_mv |
AT yousefmalik ensembleclusteringclassificationappliedtocompetingsvmandoneclassclassifiersexemplifiedbyplantmicrornasdata AT khalifawaleed ensembleclusteringclassificationappliedtocompetingsvmandoneclassclassifiersexemplifiedbyplantmicrornasdata AT abdallahloai ensembleclusteringclassificationappliedtocompetingsvmandoneclassclassifiersexemplifiedbyplantmicrornasdata |
_version_ |
1717768212778582016 |