Final BI Lab Manual

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 42

ZEAL EDUCATION SOCIETY’s

ZEAL COLLEGE OF ENGINEERING AND


RESEARCH, NARHE, PUNE

DEPARTMENT OF
COMPUTER
ENGINEERING
SEMESTER-II
[A.Y. : 2023 - 2024]

Data Science and Big Data Analytics (310251)


LABORATORY MANUAL

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 1


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 2
Department Vision and Mission

INSTITUTE To impart value added technological education through pursuit of academic excellence,
VISION research and entrepreneurial attitude.

M1: To achieve academic excellence through innovative teaching and learning process.
INSTITUTE
MISSION M2: To imbibe the research culture for addressing industry and societal needs.

M3: To provide conducive environment for building the entrepreneurial skills.

M4: To produce competent and socially responsible professionals with core human values.

DEPARTMENT To emerge as a department of repute in Computer Engineering which produces


VISION competent professionals and entrepreneurs to lead technical and betterment of
mankind.
M1: To strengthen the theoretical and practical aspects of the learning process by
teaching applications and hands on practices using modern tools and FOSS
technologies.

M2: To endavor innovative interdisciplinary research and entrepreneurship skills to


DEPARTMENT
serve the needs of Industry and Society.
MISSION
M3: To enhance industry academia dialog enabling students to inculcate professional
skills.

M4: To incorporate social and ethical awareness among the students to make them
conscientious professionals.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 3


Department
Program Educational Objectives (PEOs)

PEO1: To Impart fundamentals in science, mathematics and engineering to cater the needs of society and
Industries.

PEO2: Encourage graduates to involve in research, higher studies, and/or to become entrepreneurs.

PEO3: To Work effectively as individuals and as team members in a multidisciplinary environment with
high ethical values for the benefit of society.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 4


Savitribai Phule Pune University
BE Computer Engineering (2019 Course)
410253(C) : Business Intelligence
Teaching Scheme: Credit Examination Scheme:

PR: 02 Hours/Week 01 TW: 50 Marks

Course Objectives:

Sr.
Course Objective Statements
No.
1 To introduce the concepts and components of Business Intelligence (BI)

2 To evaluate the technologies that make up BI (data warehousing, OLAP)

3 To identify the technological architecture of BI systems·

4 To explain different data preprocessing techniques

5 To identify machine learning model as per business need

To understand the BI applications in marketing, logistics, finance and telecommunication


6 sector

a. Course Outcomes:
Sr. No. CO Statements
1 CO1: Differentiate the concepts of Decision Support System & Business Intelligence

2 CO2:Use Data Warehouse & Business Architecture to design a BI system.

3 CO3:Build graphical reports

4 CO4:Apply different data preprocessing techniques on dataset

CO5:mplement machine learning algorithms as per business needs


5

6 CO6:Identify role of BI in marketing, logistics, and finance and telecommunication sector

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 5

List of Assignments
410253(C) : Business Intelligence

Sr. Assignment Name Page


No.
1 Import the legacy data from different sources such as (Excel , Sql Server, Oracle etc.) 1
and load in the target system. ( You can download sample database such as Adventure
works, Northwind, foodmart etc.)

2 Perform the Extraction Transformation and Loading (ETL) process to construct the database in
the Sql server
3 Create the cube with suitable dimension and fact tables based on ROLAP, MOLAP and
HOLAP model
4 Import the data warehouse data in Microsoft Excel and create the Pivot table and Pivot
Chart

5 Perform the data classification using classification algorithm. Or Perform the data clustering
using clustering algorithm
Group 2
6 Mini Project: Each group of 4 Students (max) assigned one case study for this; A BI report must be
prepared outlining the following steps:
a) Problem definition, identifying which data mining task is needed.
b) Identify and use a standard data mining dataset available for the problem.

Group A

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 6


Assinment 1
Title: Import the legacy data from different sources such as (Excel , Sql Server, Oracle etc.) and load in the
target system. ( You can download sample database such as Adventure works, Northwind, foodmart etc.)

Objective of the Assignment: To introduce the concepts and components of


Business Intelligence (BI)

Prerequisite:
1. Basics of dataset extensions.
2. Concept of data import

Contents for Theory:


1. Legacy Data
2. Sources of Legacy Data
3. How to import legacy data step by step.

1.What is Legacy Data?

Legacy data, according to Business Dictionary, is "information maintained in an old or out-of-date format
or computer system that is consequently challenging to access or handle."

2. Sources of Legacy Data

Where does legacy data come from? Virtually everywhere. Figure 1 indicates that there are many sources
from which you may obtain legacy data. This includes existing databases, often relational, although non-
RDBs such as hierarchical, network, object, XML, object/relational databases, and NoSQL databases.
Files, such as XML documents or "flat files” such as configuration files and comma-delimited text files,
are also common sources of legacy data. Software, including legacy applications that have been wrapped
(perhaps via CORBA) and legacy services such as web services or CICS transactions, can also provide
access to existing information. The point to be made is that there is often far more to gaining access to
legacy data than simply writing an SQL query against an existing relational database.

lOMoARcPSD|23094708

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 7


How to import legacy data step by step.

Step 1: Open Power BI

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 8


lOMoARcPSD|23094708

Step 2: Click on Get data following list will be displayed → select Excel

Step 3: Select required file and click on Open, Navigator screen appears

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 9


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 10
lOMoARcPSD|23094708

Step 4: Select file and click on edit

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 11


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 12
lOMoARcPSD|23094708

Step 5:Power query editor appears

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 13


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 14
lOMoARcPSD|23094708

Step 6: Again, go to Get Data and select OData feed

Step 7: Paste url as http://services.odata.org/V3/Northwind/Northwind.svc/ Click


on ok

Step 8: Select orders table And click on edit


Note: If you just want to see preview you can just click on table name without clicking on checkbox Click on edit
to view table

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 15


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 16
lOMoARcPSD|23094708

Conclusion: In this way we import the Legacy datasets using the Power BI Tool.
Date:
Marks obtained:
Sign of course coordinator:
Name of course
Coordinator :

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 17


OMoARcPSD|23094708

Assignment No:2

Title: Perform the Extraction Transformation and Loading (ETL) process to construct the database in the Sql
server.

Objective of the Assignment: To introduce the concepts and components ofBusiness Intelligence
(BI)

Prerequisite:
1. Basics of ETL Tools.
2. Concept of Sql Server.

Theory:
ETL(Extract, Transform and Load)
ETL is a process in Data Warehousing and it stands for Extract, Transform and Load.
It is a process, in which an ETL tool extracts the data from various data source systems, transforms it in
the staging area and then finally, loads it into the Data Warehouse system.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 18


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 19
lOMoARcPSD|23094708

Extraction
1. Identify the Data Sources: The first step in the ETL process is to identify the data
sources. This may include files, databases, or other data repositories.
2. Extract the Data: Once the data sources are identified, we need to extract the data
from them. This may involve writing queries to extract the relevant data or using tools
such as SSIS to extract data from files or databases.
3. Validate the Data: After extracting the data, it's important to validate it to ensure
that it's accurate and complete. This may involve performing data profiling or
data quality checks.
Transformation
1. Clean and Transform the Data: The next step in the ETL process is to clean and
transform the data. This may involve removing duplicates, fixing invaliddata, or
converting data types. We can use tools such as SSIS or SQL scriptsto perform these
transformations.
2. Map the Data: Once the data is cleaned and transformed, we need to map thedata to
the appropriate tables and columns in the database. This may involvecreating a data
mapping document or using a tool such as SSIS to perform the mapping.
Loading
1. Create the Database: Before loading the data, we need to create the database and the
appropriate tables. This can be done using SQL Server Management Studio or a SQL
script.
2. Load the Data: Once the database and tables are created, we can load the data into
the database. This may involve using tools such as SSIS or writingSQL scripts to
insert the data into the appropriate tables.
3. Validate the Data: After loading the data, it's important to validate it toensure that it
was loaded correctly. This may involve performing data. profiling or data quality
checks to ensure that the data is accurate and complete.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 20


lOMoARcPSD|23094708

Perform the Extraction Transformation and Loading (ETL) process to construct thedatabase in
the SQL server.

Software requirements: SQL SERVER 2012 FULL VERSION


(SQLServer2012SPl-FullSlipstream-ENU-x86)

Steps to install SQL SERVER 2012 FULL VERSION (SQLServer2012SPl-


FullSlipstream-ENU-x86) are given in my previous post.

Step 1: Open SQL Server Managemet Studio to restore backup file

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 21


lOMoARcPSD|23094708

Step 2: Right click on Databases Restore Database

Step 3: Click on towards end of device box

Step 4: Click on Add Select path of backup files

Step 5: Select both files at a time

Step 6 : Click ok and in select backup devices window Addboth files of AdventureWorks

Step 7: Open SQL Server Data Tools


Select File New Project Business Intelligence Integration Services Project & giveappropriate
project name.
Step 8: Right click on Connection Managers in solution explorer and click onNew
Connection Manager.
Add the SSIS connection manager window.

Step 9: Select OLEDB Connection Manager and Click on Add

Step 10: Configure OLE DB Connection Manager window appears Click on New

Step 11: Select Server name(as per your machine) from drop down and databasename and click on
Test connection.
If the test connection succeeded, click on OK.

Step 12: Click on OK

Connection is added to connection manager


Step 13: Drag and drop Data Flow Task in Control Flow tab

Step 14: Drag OLE DB Source from Other Sources and drop into Data Flow tab

Step 15: Double click on OLE DB source -> OLE DB Source Editor appears->click on New to
add connection manager.
Select [Sales].[Store] table from drop down ok

Step 16: Drag ole db destination in data flow tab and connect both

Step 17: Double click on OLE DB destination


Click on New to run the query to get [OLE DB Destination] in Name of the tableor the view.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 22


lOMoARcPSD|23094708

Click on OK.

Step 18: Click on Start

Step 19: Go to SQL Server Management Studio


In database tab Adventureworks Right click on [dbo].[OLE DB Destination]Script Table
as SELECT To New Query Editor Window

Step 20: Execute the following query to get output.

USE [AdventureWorks2012]GO
SELECT [BusinessEntityID] ,
[Name]
,[SalesPersonID]
,
[Demographics] ,
[rowguid] ,
[ModifiedDate]
FROM [dbo].[OLE DB Destination]GO

Conclusion : In this way we can perform the ETL process to construct adatabase in SQL Server.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 23


15

Date:
Marks obtained:
Sign of course coordinator:
Name of course Coordinator :

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 24


lOMoARcPSD|23094708

Assignment No:3

Title of the Assignment : Create the cube with suitable dimension and fact tables based
on ROLAP, MOLAP and HOLAP model.

Objective of the Assignment : To introduce the concepts and components of Business


Intelligence (BI)

Prerequisite:

1. Basics of OLAP.
2. Concept of Multi Dimensional Cube.

Theory :

1. What is a Fact Table ?

In Business Intelligence (BI), A Fact Table is a table that stores quantitative data or facts about
a business process or activity.It is a central table in a data warehouse that provides a snapshot of
a business at a specific point in time.
For example - A Fact Table in a retail business might contain sales data for each transaction,
with dimensions such as date, product, store, and customer. Analysts can use the Fact Table to
analyze trends and patterns in sales, such as which products are selling the most, which stores
are performing well, and which customers are buying the most.

2. What is a ROLAP, MOLAP and HOLAP model

ROLAP, MOLAP, and HOLAP are three types of models used in Business Intelligence (BI)
for organizing and analyzing data:
1. ROLAP (Relational Online Analytical Processing):
In this model, data is stored in a relational database, and the analysis is performed by
joining multiple tables. ROLAP allows for complex queries and is good for handling large
amounts of data, but it may be slower due to the need for frequent joins.
2. MOLAP (Multidimensional Online Analytical Processing):
In this model, data is stored in a multidimensional database, which is optimized for fast query
performance. MOLAP is good for analyzing data in multiple dimensions, such as time,

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 25


lOMoARcPSD|23094708

geography, and product, but may be limited in its ability to handle large amounts of
data. 3. HOLAP (Hybrid Online Analytical Processing):
This model combines elements of both ROLAP and MOLAP. It stores data in both a relational
and multidimensional database, allowing for efficient analysis of both large amounts of data and
complex queries. HOLAP is a good compromise between the othertwo models, offering both
speed and flexibility.

4. Create the cube with a suitable dimension and fact tables based on OLAP ?

Step 1: Creating Data Warehouse


Let us execute our T-SQL Script to create a data warehouse with fact tables, dimensionsand
populate them with appropriate test values.
Download the T-SQL script attached with this article for creation of Sales Data Warehouse or
download from this article “Create First Data Warehouse” and run it inyour SQL Server.
Downloading "Data_WareHouse SQLScript.zip" from the article
https://www.codeproject.com/Articles/652108/Create-First-Data-WareHou se

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 26


lOMoARcPSD|23094708

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 27


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 28
lOMoARcPSD|23094708

After downloading the extract file in the folder.


Follow the given steps to run the query in SSMS (SQL Server Management Studio).
1. Open SQL Server Management Studio 2012
2. Connect Database Engine

Password for sa : admin123 (as given during installation) Click Connect.


3. Open New Query editor
4. Copy paste Scripts given below in various steps in new query editor window one by
one
5. To run the given SQL Script, press F5
6. It will create and populate “Sales_DW” database on your SQL Server OR
1. Go to the extracted sql file and double click on it.
2. New Sql Query Editor will be opened containing the Sales_DW Database.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 29


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 30
lOMoARcPSD|23094708

3. Click on execute or press F5 by selecting the query one by one or directly click on
Execute.
4. After completing execution save and close SQL Server Management studio & Reopen

to see Sales_DW in Databases Tab.

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 31


lOMoARcPSD|23094708

Step 2: Start SSDT environment and create New Data Source Go to Sql Server DataTools --
> Right click and run as administrator

Click on File → New → Project


In Business Intelligence → Analysis Services Multidimensional and Data Miningmodels →
appropriate project name → click OK

Right click on Data Sources in solution explorer → New Data Source


Data Source Wizard appears

Click on New

Select Server Name → select Use SQL Server Authentication → Select or enter adatabase
name (Sales_DW)
Note : Password for sa : admin123 (as given during installation of SQL 2012 fullversion)

Click Next

Select Inherit → Next


Click Finish

Sales_DW.ds gets created under Data Sources in Solution Explorer

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 32


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 33
lOMoARcPSD|23094708

Step 3: Creating New Data Source View


In Solution explorer right click on Data Source View → Select New Data Source View

Click Next

click Next
select FactProductSales(dbo) from Available objects and put in Includes Objects byclicking

Click Next

Click Finish

Sales DW.dsv appears in Data Source Views in Solution Explorer.

Step 4: Creating new cube


Right click on Cubes → New Cube

Select Use existing tables in Select Creation Method → Next

In Select Measure Group Tables → Select FactProductSales → Click


Next In Select Measures → check all measures → Next

In Select New Dimensions → Check all Dimensions → Next

Click on Finish

Sales_DW.cube is created

Step 5: Dimension Modification


In dimension tab → Double Click Dim Product.dim

Drag and Drop Product Name from Table in Data Source View and Add in AttributePane at
left side

Step 6: Creating Attribute Hierarchy in Date Dimension


Double click On Dim Date dimension -> Drag and Drop Fields from Table shown inData
Source View to Attributes-> Drag and Drop attributes from leftmost pane of attributes to
middle pane of Hierarchy.
Drag fields in sequence from Attributes to Hierarchy window (Year, Quarter Name,Month
22

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 34


Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 35
lOMoARcPSD|23094708

Name, Week of the Month, Full Date UK)

Step 7: Deploy Cube


Right click on Project name → Properties

This window appears

Do following changes and click on Apply & ok

Right click on project name → Deploy

Deployment successful

To process cube right click on Sales_DW.cube → Process

Click run

Browse the cube for analysis in solution explorer

Conclusion : In this way we successfully implement cube with suitable

Date:

Marks obtained:

Sign of course coordinator:

Name of course
Coordinator :
dimension and fact tables based on ROLAP, MOLAP and HOLAP model

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 36


23

Assignment No:-4
MoARcPSD|23094708

Title of the Assignment: Import the data warehouse data in Microsoft Excel and create the Pivot table
and Pivot Chart.

Objective of the Assignment: To introduce the concepts and components of Business Intelligence (BI)

Prerequisite:

1. Basics of Google Sheets.

2. Concept of Table, Chart.

Contents for Theory:

1. What is a Data Warehouse?

2. What is Pivot Table and Pivot Chart?

3. Steps for Creating a Pivot Table in Google Sheets.

4. Steps for Creating a Pivot Chart in Google Sheets.

1. What is a Data Warehouse?


A data warehouse is a centralized repository of integrated and transformed data from multiple
sources within an organization. It is designed to support business intelligence (BI) activities, such
as data analysis, reporting, and decision-making.

2. What is Pivot Table and Pivot Chart?


A pivot table is a powerful tool in spreadsheet software (such as Google Sheets or Microsoft
Excel) that allows you to summarize and analyze large datasets by grouping and summarizing
data in different ways. Pivot tables allow you to quickly create tables that show a summary of
data based on specific criteria or dimensions. For example, you can use a pivot table to
summarize sales data by region or by product category. A pivot chart is a graphical representation
of the data in a pivot table. Pivot charts allow you to visualize the summarized data in a way that
is easy to understand and interpret. They can be created based on the data in a pivot table, and can
be customized in a variety of ways to better represent the data being analyzed. Pivot charts are

24
Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 37
Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 38
lOMoARcPSD|23094708

especially useful when dealing with large amounts of data, as they can help identify patterns and
trends that might not be immediately obvious from the raw data.

3. Steps for Creating a Pivot Table in Google Sheets.


1. Open a Google Sheets document with the data you want to use for the pivot table.
2. Select the range of data you want to use for the pivot table.
3. Click on the "Data" tab in the top menu, then click on "Pivot table."
4. In the "Create Pivot Table" dialog box, select the range of data you want to use for the pivot
table and choose where you want to place the pivot table (in a new sheet or in the same sheet).
5. Click on "Create."
6. In the pivot table editor, drag and drop the columns you want to use for the pivot table into the
"Rows," "Columns," and "Values" sections.
7. To add a filter to the pivot table, drag a column into the "Filter" section.
8. To customize the values in the pivot table, click on the drop-down menu in the "Values"
section and choose the type of calculation you want to use (such as sum, count, or average). 9.
Customize any additional options in the pivot table editor (such as sorting and formatting).
10.Click on "Update" to apply the changes and create the pivot table

4. Steps for Creating a Pivot Chart in Google Sheets.


1. Open a Google Sheets document with the data you want to use for the pivot chart.
2. Select the range of data you want to use for the pivot chart.
3. Click on the "Data" tab in the top menu, then click on "Pivot table."
4. In the "Create Pivot Table" dialog box, select the range of data you want to use for the pivot
table and choose where you want to place the pivot table (in a new sheet or in the same sheet).
5. Click on "Create."
6. In the pivot table editor, drag and drop the columns you want to use for the pivot chart into the
"Rows" and "Values" sections.
7. Click on the "Chart" tab in the pivot table editor.
8. Choose the type of chart you want to use for the pivot chart from the drop-down menu.
9. Customize the chart options (such as chart title, axis labels, and colors) to your liking.
10. Click on "Update" to apply the changes and create the pivot chart.

Conclusion: In this way we pivot table and pivot chart using Google spreadsheets | Excel.

Date:
Marks obtained:
Sign of course coordinator:
Name of course Coordinator :
Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 39
D|23094708

Assignment No:-5
Title of the Assignment: Perform the data classification using classification algorithm. Or perform the
data clustering using a clustering algorithm.

Objective of the Assignment: To introduce the concepts and components of Business Intelligence (BI)

Prerequisite: 1. Basics of Tableau.

Contents for Theory:

1. What is Clustering and classification?

2. Clustering in Tableau:

3. Classification in Tableau:

Theory:

1. What is Clustering and classification?


Clustering and classification are two important techniques used in bioinformatics to analyze
biological data. Clustering is the process of grouping similar objects or data points together based on
their similarity or distance from each other. In bioinformatics, clustering is often used to group genes
or proteins based on their expression patterns or sequences. Clustering can help identify patterns and
relationships between different genes or proteins, which can provide insights into their biological
function and interactions. Classification, on the other hand, is the process of assigning a label or
category to a new observation based on its features or characteristics. In bioinformatics, classification
is often used to predict the function or activity of a new gene or protein based on its sequence or
structure. Classification can help identify new drug targets or biomarkers for disease diagnosis and
treatment. Both clustering and classification are important tools for analyzing large and complex
biological datasets and can provide valuable insights into the underlying biological processes.

Clustering in Tableau:

1. Connect to the data: Connect to the data set that you want to cluster in Tableau.

2. Drag and drop the data fields: Drag and drop the data fields into the view, and select the data points
that you want to cluster.

3. Choose a clustering algorithm: Select a clustering algorithm from the analytics pane in Tableau.
Tableau provides several built-in clustering algorithms, such as K-Means and Hierarchical
Clustering.

2
Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 40
Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 41
lOMoARcPSD|23094708

4. Define the number of clusters: Define the number of clusters that you want to create. You
can do this manually or let Tableau automatically determine the optimal number of clusters.
5. Analyze the clusters: Visualize the clusters and analyze them using Tableau's built-in
visualizations and tools.

Classification in Tableau:

1. Connect to the data: Connect to the data set that you want to classify in Tableau.

2. Drag and drop the data fields: Drag and drop the data fields into the view, and select
the target variable that you want to predict.

3. Choose a classification algorithm: Select a classification algorithm from the analytics


pane in Tableau. Tableau provides several built-in classification algorithms, such as
Decision Trees and Random Forest.

4. Define the model parameters: Define the model parameters, such as the maximum tree
depth or the number of trees to use in the forest.

5. Train the model: Train the model on a subset of the data using Tableau's built-in cross-
validation functionality.

6. Evaluate the model: Evaluate the accuracy of the model using Tableau's built-in metrics,
such as confusion matrix, precision, recall, and F1 score.

7. Predict the target variable: Use the trained model to predict the target variable for new data.

8. Visualize the results: Create visualizations to communicate the results of the classification
analysis using Tableau's built-in visualization tools.

Conclusion: In this way we implement classification and clustering using Tableau.

Date:
Marks obtained:
Sign of course
coordinator:
Name of course
Coordinator :

Department of Computer Engineering , ZCOER, Narhe, Pune-41 Page 42

You might also like