Efficient Spectral Feature Selection with Minimum Redundancy

Authors

  • Zheng Zhao Arizona State University
  • Lei Wang The Australian National University
  • Huan Liu Arizona State University

DOI:

https://doi.org/10.1609/aaai.v24i1.7671

Keywords:

feature selection, multi-output regression, L2-1 norm regularization, sparse learning, dimension reduction, machine learning, data mining

Abstract

Spectral feature selection identifies relevant features by measuring their capability of preserving sample similarity. It provides a powerful framework for both supervised and unsupervised feature selection, and has been proven to be effective in many real-world applications. One common drawback associated with most existing spectral feature selection algorithms is that they evaluate features individually and cannot identify redundant features. Since redundant features can have significant adverse effect on learning performance, it is necessary to address this limitation for spectral feature selection. To this end, we propose a novel spectral feature selection algorithm to handle feature redundancy, adopting an embedded model. The algorithm is derived from a formulation based on a sparse multi-output regression with a L2,1-norm constraint. We conduct theoretical analysis on the properties of its optimal solutions, paving the way for designing an efficient path-following solver. Extensive experiments show that the proposed algorithm can do well in both selecting relevant features and removing redundancy.

Downloads

Published

2010-07-03

How to Cite

Zhao, Z., Wang, L., & Liu, H. (2010). Efficient Spectral Feature Selection with Minimum Redundancy. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 673-678. https://doi.org/10.1609/aaai.v24i1.7671