Abstract
In an earlier paper, we introduced a new "boosting" algorithm called AdaBoost which, theoretically, can be used to significantly reduce the error of any learning algorithm that consistently generates classifiers whose performance is a little better than random guessing. We also introduced the related notion of a "pseudo-loss" which is a method for forcing a learning algorithm of multi-label conceptsto concentrate on the labels that are hardest to discriminate. In this paper, we describe experiments we carried out to assess how well AdaBoost with and without pseudo-loss, performs on real learning problems. We performed two sets of experiments. The first set compared boosting to Breiman's "bagging" method when used to aggregate various classifiers (including decision trees and single attributevalue tests). We compared the performance of the two methods on a collection of machine-learning benchmarks. In the second set of experiments, we studied in more detail the performance of boosting...
Keywords
Affiliated Institutions
Related Publications
Using output codes to boost multiclass learning problems
This paper describes a new technique for solving multiclass learning problems by combining Freund and Schapire's boosting algorithm with the main ideas of Dietterich an...
Machine Learning Bias, Statistical Bias, and Statistical Variance of Decision Tree Algorithms
The term "bias" is widely used---and with different meanings---in the fields of machine learning and statistics. This paper clarifies the uses of this term and...
Generalized Boosted Models: A guide to the gbm package
Boosting takes on various forms with different programs using different loss functions, different base models, and different optimization schemes. The gbm package takes the appr...
Bagging, boosting, and C4.S
Breiman's bagging and Freund and Schapire's boosting are recent methods for improving the predictive power of classifier learning systems. Both form a set of classifiers that ar...
A Communication-Efficient Parallel Algorithm for Decision Tree
Decision tree (and its extensions such as Gradient Boosting Decision Trees and Random Forest) is a widely used machine learning algorithm, due to its practical effectiveness and...
Publication Info
- Year
- 1996
- Type
- article
- Pages
- 148-156
- Citations
- 7561
- Access
- Closed