Large margin DragPushing strategy for centroid text categorization

作者:

Highlights:

摘要

Among all conventional methods for text categorization, centroid classifier is a simple and efficient method. However it often suffers from inductive bias (or model misfit) incurred by its assumption. DragPushing is a very simple and yet efficient method to address this so-called inductive bias problem. However, DragPushing employs only one criterion, i.e., training-set error, as its objective function that cannot guarantee the generalization capability. In this paper, we propose a generalized DragPushing strategy for centroid classifier, which we called as “Large Margin DragPushing” (LMDP). The experiments conducted on three benchmark evaluation collections show that LMDP achieved about one percent improvement over the performance of DragPushing and delivered top performance nearly as well as state-of-the-art SVM without incurring significant computational costs.

论文关键词:Text classification,Information retrieval,Machine learning

论文评审过程:Available online 4 May 2006.

论文官网地址:https://doi.org/10.1016/j.eswa.2006.04.008