Improving classifier performance through repeated sampling

作者:

Highlights:

摘要

In traditional pattern recognition, the classification decision is based on a single observation of the input. In this paper, we show that by relaxing this assumption, the performance of the classifier can be improved substantially. We present a detailed analysis of one particular method for achieving this: taking a consensus vote on the classifier's output for repeated samples of the input. We prove that this approach always yields a net improvement in recognition accuracy for common distributions of interest. Upper and lower bounds on the improvement are also discussed. Under certain conditions, it is even possible to “beat” the Bayes error bound associated with the classifier. We conclude by presenting results from three sets of experiments examining the effectiveness of the idea.

论文关键词:Bayes Risk,Classifier,Consensus sequence voting,Optical character recognition,Repeated sampling,Statistical pattern recognition

论文评审过程:Received 31 August 1995, Revised 29 October 1996, Available online 7 June 2001.

论文官网地址:https://doi.org/10.1016/S0031-3203(96)00182-3