Learning semi-lazy Bayesian network classifier under the c.i.i.d assumption

作者:

Highlights:

摘要

Bayesian network classifiers (BNCs) are powerful tools in knowledge representation and inference under conditions of uncertainty. In contrast to eager learning, lazy learning seeks to improve the classification accuracy of BNCs by forming a decision theory that is especially tailored for the testing instance, whereas it has received less attention due to the high computational cost at classification time. This study introduces the conditionally independently and identically distributed (c.i.i.d.) assumption to BNCs by assuming that all instances of the same class are conditionally independent of each other and stem from the same probability distribution. Based on this premise, we propose a novel lazy BNC, semi-lazy Bayesian network classifier (SLB), which transforms each unlabeled testing instance to a series of complete instances with discriminative supposed class labels, and then builds class-specific local BNCs for each of them. Our experimental comparison on 25 UCI datasets shows that SLB has modest training time overheads and less classification time overheads. The Friedman and Nemenyi tests show that SLB has significant zero–one loss and bias advantages over some state-of-the-art BNCs, such as selective k-dependence Bayesian classifier, k-nearest neighbor, lazy Bayesian rule and average n-dependence estimators with lazy subsumption resolution.

论文关键词:00-01,99-00,Bayesian network classifier,Eager learning,Lazy learning,Independently and identically distributed assumption

论文评审过程:Received 26 June 2020, Revised 30 July 2020, Accepted 3 September 2020, Available online 19 September 2020, Version of Record 21 September 2020.

论文官网地址:https://doi.org/10.1016/j.knosys.2020.106422