zbMATH — the first resource for mathematics

Relative cost curves: an alternative to AUC and an extension to 3-class problems. (English) Zbl 1305.93195
Summary: Performance evaluation of classifiers is a crucial step for selecting the best classifier or the best set of parameters for a classifier. Receiver Operating Characteristic (ROC) curves and Area Under the ROC Curve (AUC) are widely used to analyse performance of a classifier. However, the approach does not take into account that misclassification for different classes might have more or less serious consequences. On the other hand, it is often difficult to specify exactly the consequences or costs of misclassifications. This paper is devoted to Relative Cost Curves (RCC) – a graphical technique for visualising the performance of binary classifiers over the full range of possible relative misclassification costs. This curve provides helpful information to choose the best set of classifiers or to estimate misclassification costs if those are not known precisely. In this paper, the concept of Area Above the RCC (AAC) is introduced, a scalar measure of classifier performance under unequal misclassification costs problem. We also extend RCC to multicategory problems when misclassification costs depend only on the true class.
93E12 Identification in stochastic control theory
62H30 Classification and discrimination; cluster analysis (statistical aspects)
93C95 Application models in control theory
UCI-ml; HandTill2001
Full Text: Link
[1] Bradley, A. P.: The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognition 30 (1997), 1145-1159. · Zbl 05468651
[2] Drummond, C., Holte, R. C.: Cost curves: An improved method for visualizing classifier performance. Machine Learning 65 (2006) 95-130. · Zbl 05075646
[3] Fawcett, T.: An introduction to roc analysis. Pattern Recognition Lett. 27 (2006), 861-874.
[4] Hand, D. J.: Measuring classifier performance: a coherent alternative to the area under the ROC curve. Machine Learning 77 (2009), 103-123. · Zbl 05796832
[5] Hand, D. J., Till, R. J.: A simple generalisation of the area under the ROC curve for multiple class classification problems. Machine Learning 45 (2001), 171-186. · Zbl 1007.68180
[6] Hanley, J. A.: Receiver operating characteristic (ROC) methodology: the state of the art. Critical Reviews in Diagnostic Imaging 29 (1989), 307-335.
[7] Hernández-Orallo, J., Flach, P., Ferri, C.: Brier curves: a new cost-based visualisation of classifier performance. Proc. 28th International Conference on Machine Learning (ICML-11) (L. Getoor and T. Scheffer, ACM, New York 2011, pp. 585-592.
[8] Klawonn, F., Höppner, F., May, S.: An alternative to ROC and AUC analysis of classifiers. Advances in Intelligent Data Analysis X, (J. Gama, E. Bradley, and J. Hollmén, Springer, Berlin 2011, p. 210-221.
[9] Krzanowski, W. J., Hand, D. J.: ROC Curves for Continuous data. Chapman and Hall, London 2009. · Zbl 1288.62005
[10] Li, J., Fine, J. P.: ROC analysis with multiple classes and multiple tests: methodology and its application in microarray studies. Biostatistics 9 (2008), 566-576. · Zbl 1143.62083
[11] Murphy, P. M., Aha, D. W.: Uci repository of machine learning databases. 1992. Avaible:
[12] Mossman, D.: Three-way ROCs. Medical Decision Making 19 (1999), 78-89.
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.