Loading...
Abstract
Machine learning algorithms for inferring decision trees typically choose a single “best” tree to describe the training data. Recent research has shown that classification performance can be significantly improved by voting predictions of multiple, independently produced decision trees. This paper describes an algorithm, OB1, that makes a weighted sum over many possible models. We describe one instance of OB1, that includes all possible decision trees as well as naïve Bayesian models. OB1 is compared with a number of other decision tree and instance based learning algorithms on some of the data sets from the UCI repository. Both an information gain and an accuracy measure are used for the comparison. On the information gain measure OB1 performs significantly better than all the other algorithms. On the accuracy measure it is significantly better than all the algorithms except naïve Bayes which performs comparably to OB1.
Type
Working Paper
Type of thesis
Series
Computer Science Working Papers
Citation
Cleary, J. G. & Trigg, L. E. (1998). Experiences with a weighted decision tree learner. (Working paper 98/10). Hamilton, New Zealand: University of Waikato, Department of Computer Science.
Date
1998-05
Publisher
University of Waikato, Department of Computer Science