Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection

In this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different...

Full description

Bibliographic Details
Main Authors: Hui Luo, Jiqing Han
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8584439/
id doaj-f4e69018401b42989e1cbf3cb7ff2ee2
record_format Article
spelling doaj-f4e69018401b42989e1cbf3cb7ff2ee22021-03-29T22:08:03ZengIEEEIEEE Access2169-35362019-01-0176461647210.1109/ACCESS.2018.28889248584439Trace Ratio Criterion Based Large Margin Subspace Learning for Feature SelectionHui Luo0https://orcid.org/0000-0002-9540-8281Jiqing Han1School of Computer Science and Technology, Harbin Institute of Technology, Harbin, ChinaSchool of Computer Science and Technology, Harbin Institute of Technology, Harbin, ChinaIn this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different label and the same label, respectively. Specifically, for a given instance, the margin is the ratio of the distance of the nearest missing to that of the nearest hit rather than the difference of distances, which contributes to better balance since the distance to the nearest missing is usually much larger than the nearest hit. The proposed model seeks a subspace in which the margin metric is maximized. Moreover, considering that the nearest neighbors of a given sample are uncertain in the presence of many irrelevant features, we treat them as hidden variables and estimate the expectation of margin. To perform the feature selection, an <inline-formula> <tex-math notation="LaTeX">$\ell _{2,1}$ </tex-math></inline-formula>-norm is imposed on the subspace projection matrix to enforce row sparsity. The resulting trace ratio optimization problem, which can be connected to a nonlinear eigenvalue problem, is hard to solve. Thus, we design an efficient iterative algorithm and present a theoretical analysis of the convergence. Finally, we evaluate the proposed method by comparing it against several other state-of-the-art methods. The extensive experiments on real-world datasets show the superiority of our proposed approach.https://ieeexplore.ieee.org/document/8584439/Feature selectionreliefsubspace learningtrace ratio
collection DOAJ
language English
format Article
sources DOAJ
author Hui Luo
Jiqing Han
spellingShingle Hui Luo
Jiqing Han
Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
IEEE Access
Feature selection
relief
subspace learning
trace ratio
author_facet Hui Luo
Jiqing Han
author_sort Hui Luo
title Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
title_short Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
title_full Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
title_fullStr Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
title_full_unstemmed Trace Ratio Criterion Based Large Margin Subspace Learning for Feature Selection
title_sort trace ratio criterion based large margin subspace learning for feature selection
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description In this paper, we propose a novel feature selection model based on subspace learning with the use of a large margin principle. First, we present a new margin metric described by a given instance and its nearest missing and nearest hit, which can be explained as the nearest neighbor with a different label and the same label, respectively. Specifically, for a given instance, the margin is the ratio of the distance of the nearest missing to that of the nearest hit rather than the difference of distances, which contributes to better balance since the distance to the nearest missing is usually much larger than the nearest hit. The proposed model seeks a subspace in which the margin metric is maximized. Moreover, considering that the nearest neighbors of a given sample are uncertain in the presence of many irrelevant features, we treat them as hidden variables and estimate the expectation of margin. To perform the feature selection, an <inline-formula> <tex-math notation="LaTeX">$\ell _{2,1}$ </tex-math></inline-formula>-norm is imposed on the subspace projection matrix to enforce row sparsity. The resulting trace ratio optimization problem, which can be connected to a nonlinear eigenvalue problem, is hard to solve. Thus, we design an efficient iterative algorithm and present a theoretical analysis of the convergence. Finally, we evaluate the proposed method by comparing it against several other state-of-the-art methods. The extensive experiments on real-world datasets show the superiority of our proposed approach.
topic Feature selection
relief
subspace learning
trace ratio
url https://ieeexplore.ieee.org/document/8584439/
work_keys_str_mv AT huiluo traceratiocriterionbasedlargemarginsubspacelearningforfeatureselection
AT jiqinghan traceratiocriterionbasedlargemarginsubspacelearningforfeatureselection
_version_ 1724192229553078272