0% found this document useful (0 votes)
26 views3 pages

Exp 6

The document describes performing classification using a Naive Bayes classifier. It discusses splitting a dataset into training and test sets, fitting a Naive Bayes model to the training set, predicting results on both training and test sets, and calculating accuracy scores on both.

Uploaded by

Sayan Satpati
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
26 views3 pages

Exp 6

The document describes performing classification using a Naive Bayes classifier. It discusses splitting a dataset into training and test sets, fitting a Naive Bayes model to the training set, predicting results on both training and test sets, and calculating accuracy scores on both.

Uploaded by

Sayan Satpati
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 3

DEPARTMENT OF

COMPUTER SCIENCE&ENGINEERING

Experiment 2.2
Student Name: UID:
Branch: CSE Section: 20BCS_DM_617 B
Semester: 6th Date of Performance:
Subject Name: Data Mining Lab Subject Code: 20CSP-376

1. Aim: To perform the classification using Naïve Bayes classifier.


2. Objective: : To implement classification by Naïve Bayes classifier using
WEKA tools to build a Naïve Bayes model that can accurately classify
instances based on their attributes.

3. Script and Output:


# Naive Bayes setwd("D:\\data
mining lab") getwd()
# Importing the dataset dataset =
read.csv('Social_Network_Ads.csv') dataset
= dataset[3:5]
# Encoding the target feature as factor dataset$Purchased =
factor(dataset$Purchased, levels = c(0, 1)) # Splitting the dataset
into the Training set and Test set library(caTools)#classification -
split dataset for training and testing split =
sample.split(dataset$Purchased, SplitRatio = 0.75) training_set =
subset(dataset, split == TRUE) test_set = subset(dataset, split ==
FALSE)
# Feature Scaling
#Feature scaling is a method used to normalize the range of independent variables
or features of data training_set[-3] = scale(training_set[-3]) test_set[-3] =
scale(test_set[-3])
# Fitting Naive Baiyes Classifier to the Training set
library(e1071)
classifier = naiveBayes(x = training_set[-3], y = training_set$Purchased)
print(classifier)
DEPARTMENT OF
COMPUTER SCIENCE&ENGINEERING

# Predicting train set results y_pred_train =


predict(classifier, newdata = training_set[-3])
# Making the Confusion Matrix for training set
cm_train = table(training_set[, 3], y_pred_train)
print(cm_train)
#Accuracy on training data accuracy_train <-
sum(diag(cm_train))/sum(cm_train) cat("\nAccuracy
on training set: ", accuracy_train)
# Predicting the Test set results y_pred_test =
predict(classifier, newdata = test_set[-3]) # Making the
Confusion Matrix for testing set cm_test =
table(test_set[, 3], y_pred_test) print(cm_test) #
Accuracy on test data accuracy_test <-
sum(diag(cm_test))/sum(cm_test) cat("\nAccuracy on
test set: ", accuracy_test)
DEPARTMENT OF
COMPUTER SCIENCE&ENGINEERING

You might also like