AI
]
Bayes Naive Classifiers
There is an important disctinction between generative and discriminative models. Given input data x and you want to classify the data into labels y. A generative model learns the joint probability distribution p(x,y) and a discriminative model learns the conditional probability distribution p(y|x) - which you should read as “the probability of y given x”(original answer).
Why Naive?
We assume that every feature is independent of the other ones.
Why Bayes?
We apply Bayes’ theorem in the classfier’s decision rule. Naive Bayes can be trained very efficiently. With a single pass over the training data, it computes the conditional probability distribution of each feature given each label. For prediction, it applies Bayes’ theorem to compute the conditional probability distribution of each label given an observation.
Why Classifiers?
NBC classify, meaning it predicts which set of categories a new observation belongs.
Naive Bayes classifiers are a family of probabilistic classifiers that is able to predict, given an observation of input, a probability distribution over a set of classes. Given a problem instance to be classified, represented by a vector \({\displaystyle \mathbf {x} =(x_{1},\ldots ,x_{n})}{\displaystyle \mathbf {x} =(x_{1},\ldots ,x_{n})}\) representing some n features (independent variables), it assigns to this instance probabilities
{\displaystyle p(C_{k}\mid x_{1},\ldots ,x_{n})\,}{\displaystyle p(C_{k}\mid x_{1},\ldots ,x_{n})\,} for each of K possible outcomes or classes {\displaystyle C_{k}}C_{k}
Its pros and cons can be summarized as follows Pros:
- Highly scalable, requiring a number of parameters linear in the number of predictors in a learning problem and hence faster compared to complicated algorithms.
- Works well for high dimensional problems such as text classification because it makes a very strong assumption.
Cons:
- The naive assumptionis not usually the case in real life problems.
Construct Naive Bayers
References: