

Author: Zhang Harry Su Jiang
Publisher: Taylor & Francis Ltd
ISSN: 1362-3079
Source: Journal of Experimental & Theoretical Artificial Intelligence, Vol.20, Iss.2, 2008-06, pp. : 79-93
Disclaimer: Any content in publications that violate the sovereignty, the constitution or regulations of the PRC is not accepted or approved by CNPIEC.
Abstract
It is well known that naive Bayes performs surprisingly well in classification, but its probability estimation is poor. AUC (the area under the receiver operating characteristics curve) is a measure different from classification accuracy and probability estimation, which is often used to measure the quality of rankings. Indeed, an accurate ranking of examples is often more desirable than a mere classification. What is the general performance of naive Bayes in yielding optimal ranking, measured by AUC? In this paper, we study it systematically by both empirical experiments and theoretical analysis. In our experiments, we compare naive Bayes with a state-of-the-art decision-tree learning algorithm C4.4 for ranking, and some popular extensions of naive Bayes which achieve a significant improvement over naive Bayes in classification, such as the selective Bayesian classifier (SBC) and tree-augmented naive Bayes (TAN). Our experimental results show that naive Bayes performs significantly better than C4.4 and comparably with TAN. This provides empirical evidence that naive Bayes performs well in ranking. Then we analyse theoretically the optimality of naive Bayes in ranking. We study two example problems: conjunctive concepts and m-of-n concepts, which have been used in analysing the performance of naive Bayes in classification. Surprisingly, naive Bayes performs optimally on them in ranking, even though it does not in classification. We present and prove a sufficient condition for the optimality of naive Bayes in ranking. From both empirical and theoretical studies, we believe that naive Bayes is a competitive model for ranking.
Related content


NUMERIC MAPPING AND LEARNABILITY OF NAIVE BAYES
By ZHANG HARRY
Applied Artificial Intelligence, Vol. 17, Iss. 5-6, 2003-05 ,pp. :


Technical Note: Naive Bayes for Regression
By Frank E.
Machine Learning, Vol. 41, Iss. 1, 2000-10 ,pp. :


Mixed-membership naive Bayes models
By Shan Hanhuai
Data Mining and Knowledge Discovery, Vol. 23, Iss. 1, 2011-07 ,pp. :




Not So Naive Bayes: Aggregating One-Dependence Estimators
Machine Learning, Vol. 58, Iss. 1, 2005-01 ,pp. :