Abstract
Feature selection is an important problem for pattern classification systems. We study how to select good features according to the maximal statistical dependency criterion based on mutual information. Because of the difficulty in directly implementing the maximal dependency condition, we first derive an equivalent form, called minimal-redundancy-maximal-relevance criterion (mRMR), for first-order incremental feature selection. Then, we present a two-stage feature selection algorithm by combining mRMR and other more sophisticated feature selectors (e.g., wrappers). This allows us to select a compact set of superior features at very low cost. We perform extensive experimental comparison of our algorithm and other methods using three different classifiers (naive Bayes, support vector machine, and linear discriminate analysis) and four different data sets (handwritten digits, arrhythmia, NCI cancer cell lines, and lymphoma tissues). The results confirm that mRMR leads to promising improvement on feature selection and classification accuracy.
Keywords
Affiliated Institutions
Related Publications
Input feature selection by mutual information based on Parzen window
Mutual information is a good indicator of relevance between variables, and have been used as a measure in several feature selection algorithms: However, calculating the mutual i...
Biomarker Identification by Feature Wrappers
Gene expression studies bridge the gap between DNA information and trait information by dissecting biochemical pathways into intermediate components between genotype and phenoty...
Molecular Similarity Searching Using Atom Environments, Information-Based Feature Selection, and a Naïve Bayesian Classifier
A novel technique for similarity searching is introduced. Molecules are represented by atom environments, which are fed into an information-gain-based feature selection. A naïve...
Minimum redundancy feature selection from microarray gene expression data
Selecting a small subset of genes out of the thousands of genes in microarray data is important for accurate classification of phenotypes. Widely used methods typically rank gen...
Feature selection for multiclass discrimination via mixed-integer linear programming
We reformulate branch-and-bound feature selection employing L/sub /spl infin// or particular L/sub p/ metrics, as mixed-integer linear programming (MILP) problems, affording con...
Publication Info
- Year
- 2005
- Type
- article
- Volume
- 27
- Issue
- 8
- Pages
- 1226-1238
- Citations
- 10050
- Access
- Closed
External Links
Social Impact
Social media, news, blog, policy document mentions
Citation Metrics
Cite This
Identifiers
- DOI
- 10.1109/tpami.2005.159