Classifier variability: Accounting for training and testing

作者:

Highlights:

摘要

We categorize the statistical assessment of classifiers into three levels: assessing the classification performance and its testing variability conditional on a fixed training set, assessing the performance and its variability that accounts for both training and testing, and assessing the performance averaging over training sets and its variability that accounts for both training and testing. We derived analytical expressions for the variance of the estimated AUC and provide freely available software implemented with an efficient computation algorithm. Our approach can be applied to assess any classifier that has ordinal (continuous or discrete) outputs. Applications to simulated and real datasets are presented to illustrate our methods.

论文关键词:Classifier evaluation,Training variability,Classifier stability,U-statistics,AUC

论文评审过程:Received 24 September 2011, Accepted 30 December 2011, Available online 11 January 2012.

论文官网地址:https://doi.org/10.1016/j.patcog.2011.12.024