Abstract
In this paper, we introduce a new method to improve the performance of combining boosting and naïve Bayesian. Instead of combining boosting and Naïve Bayesian learning directly, which was proved to be unsatisfactory to improve performance, we select the training samples dynamically by bootstrap method for the construction of naïve Bayesian classifiers, and hence generate very different or unstable base classifiers for boosting. Besides, we devise a modification for the weight adjusting of boosting algorithm in order to achieve this goal: minimizing the overlapping errors of its constituent classifiers. We conducted series of experiments, which show that the new method not only has performance much better than naïve Bayesian classifiers or directly boosted naïve Bayesian ones, but also much quicker to obtain optimal performance than boosting stumps and boosting decision trees incorporated with naïve Bayesian learning.
Supported by the National Grand Fundamental Research 973 Program of China under Grant No.G1998030414 and the National Natural Science Foundation of China under Grant No.79990580.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Freund, Y., Schapire, R.: A Decision-theoretic Generalization of On-line Learning and an Application to Boosting. Journal of Computer and System Sciences (1997), 55(1), 119–139
Ting, K., Zheng, Z.: Improving the Performance of Boosting for Naïve Bayesian Classification. School of Computing and Mathematics, Deakin University (2000)
Breiman, L.: Bagging Predictors. Machine Learning (1996), 24(2), 123–140
Bremain, L.: Bias, Variance, and Arcing Classifiers. Machine Learning (2000)
Quinlan, J.R.: Mini-Boosting Decision Trees. AI Access Foundation and Morgan Kaufmann Publishers (1998)
Dietterich, T.G.: Machine Learning Research. AI Magazine (1997), 18(4), 97–136
McCallum, A., Nigam, K.: A Comparison of Event Models for Naive Bayesian Text Classification. Just Research. 4616 Henry Street Pittsburgh, PA 15213 (1999)
Schapire, R., Singer, Y.: BoosTexter: A Boosting-based System for Text Categorization. Machine Learning (2000), 39(2/3), 135–168
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Diao, L., Hu, K., Lu, Y., Shi, C. (2002). A Method to Boost Naïve Bayesian Classifiers. In: Chen, MS., Yu, P.S., Liu, B. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2002. Lecture Notes in Computer Science(), vol 2336. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-47887-6_11
Download citation
DOI: https://doi.org/10.1007/3-540-47887-6_11
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-43704-8
Online ISBN: 978-3-540-47887-4
eBook Packages: Springer Book Archive