Loading...
Thumbnail Image
Item

An entropy gain measure of numeric prediction performance

Abstract
Categorical classifier performance is typically evaluated with respect to error rate, expressed as a percentage of test instances that were not correctly classified. When a classifier produces multiple classifications for a test instance, the prediction is counted as incorrect (even if the correct class was one of the predictions). Although commonly used in the literature, error rate is a coarse measure of classifier performance, as it is based only on a single prediction offered for a test instance. Since many classifiers can produce a class distribution as a prediction, we should use this to provide a better measure of how much information the classifier is extracting from the domain. Numeric classifiers are a relatively new development in machine learning, and as such there is no single performance measure that has become standard. Typically these machine learning schemes predict a single real number for each test instance, and the error between the predicted and actual value is used to calculate a myriad of performance measures such as correlation coefficient, root mean squared error, mean absolute error, relative absolute error, and root relative squared error. With so many performance measures it is difficult to establish an overall performance evaluation. The next section describes a performance measure for machine learning schemes that attempts to overcome the problems with current measures. In addition, the same evaluation measure is used for categorical and numeric classifier.
Type
Working Paper
Type of thesis
Series
Computer Science Working Papers
Citation
Trigg, L. (1998). An entropy gain measure of numeric prediction performance. (Working paper 98/11). Hamilton, New Zealand: University of Waikato, Department of Computer Science.
Date
1998-05
Publisher
University of Waikato, Department of Computer Science
Degree
Supervisors
Rights