Naive Bayes
Sources:
Overview
It assumes conditional independence between every pair of features given the value of the class. This is why this method is called "naive".
The Bayes theorem states:
P(y∣x1,...,xn)=P(x1,...,xn)P(y)P(x1,...,xn∣y)
When assuming conditional independence:
P(y∣x1,...,xn)=P(x1,...,xn)P(y)P(x1,...,xn∣y)
So, we can use the following classification rule:
y^=argymaxP(y)i=1∏nP(xi∣y)
Advantages
NB classifiers have worked quite well in many real-world situations, famously document classification, sentiment analysis and spam filtering.
They require a small amount of training data to estimate the necessary parameters.
They can be extremely fast compared to more sophisticated methods.
Disadvantages
Although NB is considered as a decent classifier, it is known to be a bad estimator. So their output probabilities are not to be taken too seriously.
Types of NB classifiers
Last updated