Multinomial naive Bayes for text categorization revisited
Citation
Export citationKibriya, A. M., Frank, E., Pfahringer, B. & Holmes, G. (2005). Multinomial naive Bayes for text categorization revisited. In G.I. Webb & Xinghuo Yu(Eds.), Proceedings of 17th Australian Joint Conference on Artificial Intelligence, Cairns, Australia, December 4-6, 2004.(pp. 488-499). Berlin: Springer.
Permanent Research Commons link: https://hdl.handle.net/10289/1448
Abstract
This paper presents empirical results for several versions of the multinomial naive Bayes classifier on four text categorization problems, and a way of improving it using locally weighted learning. More specifically, it compares standard multinomial naive Bayes to the recently proposed transformed weight-normalized complement naive Bayes classifier (TWCNB) [1], and shows that some of the modifications included in TWCNB may not be necessary to achieve optimum performance on some datasets. However, it does show that TFIDF conversion and document length normalization are important. It also shows that support vector machines can, in fact, sometimes very significantly outperform both methods. Finally, it shows how the performance of multinomial naive Bayes can be improved using locally weighted learning. However, the overall conclusion of our paper is that support vector machines are still the method of choice if the aim is to maximize accuracy.
Date
2005Publisher
Springer