Abstract

For many types of machine learning algorithms, one can compute the statistically `optimal' way to select training data. In this paper, we review how optimal data selection techniques have been used with feedforward neural networks. We then show how the same principles may be used to select data for two alternative, statistically-based learning architectures: mixtures of Gaussians and locally weighted regression. While the techniques for neural networks are computationally expensive and approximate, the techniques for mixtures of Gaussians and locally weighted regression are both efficient and accurate. Empirically, we observe that the optimality criterion sharply decreases the number of training examples the learner needs in order to achieve good performance.

Keywords

Computer scienceMachine learningArtificial intelligenceArtificial neural networkRegressionSelection (genetic algorithm)Feedforward neural networkTraining setMathematicsStatistics

Affiliated Institutions

Related Publications

Publication Info

Year
1996
Type
article
Volume
4
Pages
129-145
Citations
1241
Access
Closed

External Links

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

1241
OpenAlex

Cite This

David Cohn, Zoubin Ghahramani, Michael I. Jordan (1996). Active Learning with Statistical Models. Journal of Artificial Intelligence Research , 4 , 129-145. https://doi.org/10.1613/jair.295

Identifiers

DOI
10.1613/jair.295