Leathart, T., Frank, E., Holmes, G., & Pfahringer, B. (2017). Probability calibration trees. In M.-L. Zhang & Y.-K. Noh (Eds.), Proceedings of 9th Asian Conference on Machine Learning (Vol. PMLR 77, pp. 145–160). Seoul, Korea.
Permanent Research Commons link: https://hdl.handle.net/10289/11515
Obtaining accurate and well calibrated probability estimates from classifiers is useful in many applications, for example, when minimising the expected cost of classifications. Existing methods of calibrating probability estimates are applied globally, ignoring the potential for improvements by applying a more fine-grained model. We propose probability calibration trees, a modification of logistic model trees that identifies regions of the input space in which different probability calibration models are learned to improve performance. We compare probability calibration trees to two widely used calibration methods—isotonic regression and Platt scaling—and show that our method results in lower root mean squared error on average than both methods, for estimates produced by a variety of base learners.
© 2017 T. Leathart, E. Frank, G. Holmes & B. Pfahringer.