Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
In this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2019-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8584439/ |
id |
doaj-f4e69018401b42989e1cbf3cb7ff2ee2 |
---|---|
record_format |
Article |
spelling |
doaj-f4e69018401b42989e1cbf3cb7ff2ee22021-03-29T22:08:03ZengIEEEIEEE Access2169-35362019-01-0176461647210.1109/ACCESS.2018.28889248584439Trace Ratio Criterion Based Large Margin Subspace Learning for Feature SelectionHui Luo0https://orcid.org/0000-0002-9540-8281Jiqing Han1School of Computer Science and Technology, Harbin Institute of Technology, Harbin, ChinaSchool of Computer Science and Technology, Harbin Institute of Technology, Harbin, ChinaIn this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different label and the same label, respectively. Specifically, for a given instance, the margin is the ratio of the distance of the nearest missing to that of the nearest hit rather than the difference of distances, which contributes to better balance since the distance to the nearest missing is usually much larger than the nearest hit. The proposed model seeks a subspace in which the margin metric is maximized. Moreover, considering that the nearest neighbors of a given sample are uncertain in the presence of many irrelevant features, we treat them as hidden variables and estimate the expectation of margin. To perform the feature selection, an <inline-formula> <tex-math notation="LaTeX">$\ell _{2,1}$ </tex-math></inline-formula>-norm is imposed on the subspace projection matrix to enforce row sparsity. The resulting trace ratio optimization problem, which can be connected to a nonlinear eigenvalue problem, is hard to solve. Thus, we design an efficient iterative algorithm and present a theoretical analysis of the convergence. Finally, we evaluate the proposed method by comparing it against several other state-of-the-art methods. The extensive experiments on real-world datasets show the superiority of our proposed approach.https://ieeexplore.ieee.org/document/8584439/Feature selectionreliefsubspace learningtrace ratio |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Hui Luo Jiqing Han |
spellingShingle |
Hui Luo Jiqing Han Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection IEEE Access Feature selection relief subspace learning trace ratio |
author_facet |
Hui Luo Jiqing Han |
author_sort |
Hui Luo |
title |
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection |
title_short |
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection |
title_full |
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection |
title_fullStr |
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection |
title_full_unstemmed |
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection |
title_sort |
trace ratio criterion based large margin subspace learning for feature selection |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2019-01-01 |
description |
In this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different label and the same label, respectively. Specifically, for a given instance, the margin is the ratio of the distance of the nearest missing to that of the nearest hit rather than the difference of distances, which contributes to better balance since the distance to the nearest missing is usually much larger than the nearest hit. The proposed model seeks a subspace in which the margin metric is maximized. Moreover, considering that the nearest neighbors of a given sample are uncertain in the presence of many irrelevant features, we treat them as hidden variables and estimate the expectation of margin. To perform the feature selection, an <inline-formula> <tex-math notation="LaTeX">$\ell _{2,1}$ </tex-math></inline-formula>-norm is imposed on the subspace projection matrix to enforce row sparsity. The resulting trace ratio optimization problem, which can be connected to a nonlinear eigenvalue problem, is hard to solve. Thus, we design an efficient iterative algorithm and present a theoretical analysis of the convergence. Finally, we evaluate the proposed method by comparing it against several other state-of-the-art methods. The extensive experiments on real-world datasets show the superiority of our proposed approach. |
topic |
Feature selection relief subspace learning trace ratio |
url |
https://ieeexplore.ieee.org/document/8584439/ |
work_keys_str_mv |
AT huiluo traceratiocriterionbasedlargemarginsubspacelearningforfeatureselection AT jiqinghan traceratiocriterionbasedlargemarginsubspacelearningforfeatureselection |
_version_ |
1724192229553078272 |