Abstract
Research in machine learning, statistics and related fields has produced a wide variety of algorithms for classification. However, most of these algorithms assume that all errors have the same cost, which is seldom the case in KDD problems. Individually making each classification learner costsensitive is laborious, and often non-trivial. In this paper we propose a principled method for making an arbitrary classifier cost-sensitive by wrapping a cost-minimizing procedure around it. This procedure, called MetaCost, treats the underlying classifier as a black box, requiring no knowledge of its functioning or change to it. Unlike stratification, MetaCost, is applicable to any number of classes and to arbitrary cost matrices. Empirical trials on a large suite of benchmark databases show that MetaCost almost always produces large cost reductions compared to the cost-blind classifier used (C4.5RULES) and to two forms of stratification. Further tests identify the key components of MetaCost and those that can be varied without substantial loss. Experiments on a larger database indicate that MetaCost scales well.
Keywords
Affiliated Institutions
Related Publications
Quickly Boosting Decision Trees - Pruning Underachieving Features Early
Boosted decision trees are one of the most popular and successful learning techniques used today. While exhibiting fast speeds at test time, relatively slow training makes them ...
Feature Engineering and Classifier Ensemble for KDD Cup 2010
KDD Cup 2010 is an educational data mining competition. Participants are asked to learn a model from students ’ past behavior and then predict their future performance. At Natio...
Cost-sensitive learning by cost-proportionate example weighting
We propose and evaluate a family of methods for converting classifier learning algorithms and classification theory into cost-sensitive algorithms and theory. The proposed conve...
Approximate Splitting for Ensembles of Trees using Histograms
Recent work in classification indicates that significant improvements in accuracy can be obtained by growing an ensemble of classifiers and having them vote for the most popular...
Building projectable classifiers of arbitrary complexity
Conventional methods for classifier design often suffer from having two conflicting goals-to develop arbitrarily complex decision boundaries to suit a given problem, and at the ...
Publication Info
- Year
- 1999
- Type
- article
- Pages
- 155-164
- Citations
- 1289
- Access
- Closed
External Links
Social Impact
Social media, news, blog, policy document mentions
Citation Metrics
Cite This
Identifiers
- DOI
- 10.1145/312129.312220