Quadratic boosting

作者:

Highlights:

摘要

This paper presents a strategy to improve the AdaBoost algorithm with a quadratic combination of base classifiers. We observe that learning this combination is necessary to get better performance and is possible by constructing an intermediate learner operating on the combined linear and quadratic terms. This is not trivial, as the parameters of the base classifiers are not under direct control, obstructing the application of direct optimization. We propose a new method realizing iterative optimization indirectly. First we train a classifier by randomizing the labels of training examples. Subsequently, the input learner is called repeatedly with a systematic update of the labels of the training examples in each round. We show that the quadratic boosting algorithm converges under the condition that the given base learner minimizes the empirical error. We also give an upper bound on the VC-dimension of the new classifier. Our experimental results on 23 standard problems show that quadratic boosting compares favorably with AdaBoost on large data sets at the cost of training speed. The classification time of the two algorithms, however, is equivalent.

论文关键词:AdaBoost,Boosting algorithm,Coordinate descent,Generalization error,Object detection,Quadratic boosting,Randomized relabeling,VC-dimension

论文评审过程:Received 31 May 2006, Revised 1 May 2007, Accepted 6 May 2007, Available online 24 May 2007.

论文官网地址:https://doi.org/10.1016/j.patcog.2007.05.008