0% found this document useful (0 votes)
23 views7 pages

Naive Bayes Classifier Overview

Uploaded by

writers.way01
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views7 pages

Naive Bayes Classifier Overview

Uploaded by

writers.way01
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Naive Bayes Classifier Overview

Naive Bayes Classifiers are a powerful and simple probabilistic classification technique based on
Bayes' Theorem with an assumption of independence between predictors. Here are some key
points and concepts you might want to cover:

1. Bayes' Theorem:

Explain how this theorem is used to calculate the probability of a class given a set of
features.
2. Assumption of Independence: Discuss the 'naive' assumption that all features are
independent given the class label, which simplifies the calculation of the likelihood.
3. Types of Naive Bayes Classifiers:
o Gaussian Naive Bayes: Assumes that the continuous features follow a Gaussian
(normal) distribution.
o Multinomial Naive Bayes: Useful for discrete features (e.g., word counts in text
classification).
o Bernoulli Naive Bayes: Used for binary/boolean features.
4. Application: Discuss common applications like text classification (spam detection),
sentiment analysis, and medical diagnosis.
5. Example Calculation: Provide a simple example to calculate the posterior probability for
a given class. For instance, classifying a document as spam or not spam based on word
frequencies.
6. Advantages and Disadvantages:
o Advantages: Simple, fast, works well with high-dimensional data.
o Disadvantages: The assumption of feature independence is often unrealistic.

General Derivation for Naive Bayes Classifier

Bayes' Theorem

Bayes' Theorem provides a way to update the probability estimate for a hypothesis as more
evidence is acquired. The theorem is given by:
Applying Bayes' Theorem to Classification

In the context of Naive Bayes classification, we are interested in finding the probability of a
class C (e.g., Spam or Not Spam) given a set of features (words) 𝑊 = {𝑤1 , 𝑤1 , 𝑤1 , ⋯ , 𝑤𝑛 }

Naive Bayes Assumption

The Naive Bayes classifier assumes that the features (words) are conditionally independent
given the class. This simplifies the calculation of the likelihood:
Let’s go through an example of how to calculate the intermediate probabilities for a Naive Bayes
classifier in a spam detection task. We'll assume we have a small dataset with some emails labeled as
"spam" or "not spam."

Example Dataset

Let's consider the following dataset of emails with words and their corresponding labels:

Count the occurrences of each word in spam and not spam emails:
To avoid zero probabilities, we'll use Laplace smoothing (add 1 to each count and add the
number of unique words to the denominator).

Number of unique words: 10

Total words in Spam emails: 2 + 2 + 1 + 1 + 2 = 8

Total words in Not Spam emails: 1 + 1 + 1 + 1 + 1 = 5

You might also like