0% found this document useful (0 votes)
58 views

Module 4 - Study Material - Overview of Predictive Analytics

Data mining encompasses techniques for discovering patterns in large datasets, including association rule mining, classification, clustering, regression analysis, and time series analysis. Supervised learning uses labeled data to learn input-output mappings, while unsupervised learning explores inherent structures in unlabeled data. Classification categorizes inputs into classes, and prediction estimates continuous numerical outputs. Typical steps in data mining are data collection, preprocessing, exploratory analysis, model building, evaluation, and deployment.

Uploaded by

Anand Panduranga
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
58 views

Module 4 - Study Material - Overview of Predictive Analytics

Data mining encompasses techniques for discovering patterns in large datasets, including association rule mining, classification, clustering, regression analysis, and time series analysis. Supervised learning uses labeled data to learn input-output mappings, while unsupervised learning explores inherent structures in unlabeled data. Classification categorizes inputs into classes, and prediction estimates continuous numerical outputs. Typical steps in data mining are data collection, preprocessing, exploratory analysis, model building, evaluation, and deployment.

Uploaded by

Anand Panduranga
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

GLOBAL ACADEMY OF TE CHNOLOGY,

BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

Predictive and Time Series Analysis

Module 4: Overview of Predictive Analytics

Core ideas in Data Mining:

Data mining encompasses various techniques and methods aimed at discovering patterns, trends, and
insights from large datasets. Some core ideas in data mining:

1. Association Rule Mining: This involves discovering relationships or associations between


variables in large datasets. It's often used in market basket analysis to uncover patterns like
"Customers who buy X also tend to buy Y."

2. Classification: Classification is about categorizing data into predefined classes or categories


based on their characteristics. It's used for tasks like spam email detection, sentiment analysis,
and medical diagnosis.

3. Clustering: Clustering is the process of grouping similar data points together based on certain
features or attributes. It's useful for segmentation in marketing, anomaly detection, and
exploratory data analysis.

4. Regression Analysis: Regression analysis explores the relationship between a dependent


variable and one or more independent variables. It's used to predict numerical values and
understand the strength of relationships between variables.

5. Time Series Analysis: Time series analysis deals with data points collected or recorded over a
period of time. It's used for forecasting future trends, detecting seasonality, and understanding
underlying patterns in time-dependent data.

6. Anomaly Detection: Anomaly detection focuses on identifying unusual patterns or outliers in


data that deviate from normal behavior. It's applied in fraud detection, network security, and
fault detection in industrial processes.

7. Dimensionality Reduction: Dimensionality reduction techniques aim to reduce the number of


variables in a dataset while preserving its essential features. This helps in visualizing high-
dimensional data and improving the efficiency of machine learning algorithms.

8. Text Mining: Text mining involves extracting valuable insights and patterns from unstructured
textual data. It includes tasks like sentiment analysis, topic modeling, and named entity
recognition.

9. Feature Selection and Extraction: Feature selection involves selecting the most relevant
features or variables that contribute the most to the predictive model's performance. Feature
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

extraction transforms raw data into a more compact representation by capturing essential
information.

10. Data Preprocessing: Data preprocessing involves cleaning, transforming, and preparing raw data
for analysis. It includes tasks like handling missing values, normalization, and encoding
categorical variables.

These core ideas in data mining form the foundation for various applications across industries, helping
organizations make data-driven decisions and gain valuable insights from their datasets.

Supervised vs Unsupervised Learning:

Supervised and unsupervised learning are two primary approaches in machine learning, each serving
different purposes and requiring different types of data.

1. Supervised Learning:

 In supervised learning, the algorithm is trained on a labeled dataset, meaning each input
is associated with a corresponding target output.

 The goal is to learn a mapping function from input variables to output variables.

 During training, the algorithm learns from the labeled data by adjusting its parameters
to minimize the difference between predicted and actual outputs.

 Once trained, the model can make predictions on unseen data by generalizing patterns
learned during training.

 Examples:

 Classification: Predicting whether an email is spam or not based on its content.


Here, the input (email content) is associated with a label (spam or not spam).

 Regression: Predicting house prices based on features like square footage,


number of bedrooms, etc. Here, the input (features of the house) is associated
with a numerical target (price).

2. Unsupervised Learning:

 In unsupervised learning, the algorithm is trained on an unlabeled dataset, meaning


there are no predefined target outputs.

 The goal is to find hidden patterns, structures, or relationships in the data without
explicit guidance.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

 The algorithm explores the data, identifying similarities, differences, and natural
groupings among data points.

 Since there are no labels, the evaluation of an unsupervised learning model is often
subjective and relies on domain knowledge or downstream tasks.

 Examples:

 Clustering: Grouping similar customers based on their purchasing behavior


without any prior labels. The algorithm identifies clusters of customers with
similar preferences.

 Dimensionality Reduction: Reducing the number of features while preserving


essential information. Techniques like Principal Component Analysis (PCA) help
in visualizing high-dimensional data and uncovering underlying patterns.

In summary, supervised learning requires labeled data and aims to learn the relationship between inputs
and outputs, while unsupervised learning explores the data's inherent structure without explicit
guidance, often to discover hidden patterns or groupings.

Classification vs Prediction:

Classification and prediction are both tasks commonly performed in supervised learning, but they serve
different purposes and involve different types of outputs. Here's a breakdown of each, along with
examples:

1. Classification:

- Purpose: Classification is the task of categorizing input data into predefined classes or categories.

- Output: The output of a classification model is a discrete class label.

- Example: Spam Email Detection

- In this example, the task is to classify emails as either "spam" or "not spam" based on their content
and features.

- Input: Email content, sender information, subject line, etc.

- Output: Binary classification - "spam" or "not spam."

- The model is trained on a dataset where each email is labeled as spam or not spam. It learns to
differentiate between spam and legitimate emails based on features extracted from the email content.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

- Once trained, the model can classify new, unseen emails as either spam or not spam based on the
patterns learned during training.

2. Prediction:

- Purpose: Prediction, also known as regression, involves estimating a continuous numerical value
based on input variables.

- Output: The output of a prediction model is a numerical value.

- Example: House Price Prediction

- In this example, the task is to predict the price of a house based on its features such as square
footage, number of bedrooms, location, etc.

- Input: Features of the house - square footage, number of bedrooms, location, etc.

- Output: Predicted house price (a continuous numerical value).

- The model is trained on a dataset where each data point represents a house with its features and
corresponding sale price.

- The model learns the relationship between the input features and the house prices, allowing it to
predict the price of a new house based on its features.

- The output of the model is a numerical value representing the predicted price of the house.

In summary, classification involves categorizing input data into discrete classes or categories, while
prediction (regression) involves estimating a continuous numerical value based on input variables. Both
tasks are examples of supervised learning but differ in the type of output they produce.

Steps in Data Mining:

Data mining involves several steps to extract useful patterns and insights from large datasets. Here's a
brief overview of the typical steps involved in data mining, along with an example:

1. Data Collection:

- Gather relevant data from various sources such as databases, data warehouses, or external sources
like web scraping.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

- Example: Collecting sales data from a retail store, including customer demographics, purchase
history, and product information.

2. Data Preprocessing:

- Clean the data by handling missing values, removing duplicates, and dealing with outliers.

- Transform the data into a suitable format for analysis, including normalization or standardization.

- Example: Removing duplicate entries from the sales data and filling in missing values in customer
demographic information.

3. Exploratory Data Analysis (EDA):

- Explore the dataset to understand its characteristics, distributions, and relationships between
variables.

- Visualize the data using techniques like histograms, scatter plots, and correlation matrices.

- Example: Plotting histograms of product sales to understand their distribution and identifying
correlations between different product categories.

4. Feature Selection/Extraction:

- Identify the most relevant features (variables) that contribute to the analysis.

- Perform feature extraction techniques to create new meaningful features from existing ones.

- Example: Selecting the most important features such as customer age, income, and purchase
frequency for predicting customer churn.

5. Model Building:

- Choose appropriate data mining algorithms based on the nature of the problem (classification,
regression, clustering, etc.).

- Train the selected models on the prepared dataset.

- Example: Building a classification model using decision trees to predict whether a customer will
purchase a particular product based on their demographic information and past purchase history.

6. Model Evaluation:

- Assess the performance of the models using evaluation metrics such as accuracy, precision, recall, or
F1-score.

- Validate the models using techniques like cross-validation to ensure generalizability.


GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

- Example: Evaluating the accuracy of the classification model by comparing its predictions against
actual purchase behavior in a test dataset.

7. Model Deployment:

- Deploy the trained model into production for making predictions on new, unseen data.

- Integrate the model into existing systems or applications for real-time decision-making.

- Example: Implementing the classification model within a customer relationship management (CRM)
system to identify potential product recommendations for individual customers during their online
shopping experience.

These steps provide a structured approach to extract valuable insights and patterns from data, enabling
informed decision-making and driving business value.

SEMMA Approach (Sample, Explore, Modify, Model, Assess):

The SEMMA approach, developed by SAS Institute, is a structured methodology for data mining that
consists of five key stages: Sample, Explore, Modify, Model, and Assess. Here's a brief explanation of
each stage along with an example:

1. Sample:

- In the sample stage, a representative subset of the data is selected for analysis.

- This subset should be large enough to capture the characteristics of the entire dataset but small
enough to be manageable.

- Example: In a marketing campaign analysis, a sample of customer data, including demographics,


purchase history, and response to previous campaigns, is selected for further analysis.

2. Explore:

- In the explore stage, the selected data is explored to gain insights and identify patterns.

- Techniques such as data visualization, summary statistics, and correlation analysis are used to
understand the relationships between variables.

- Example: Visualizing customer purchase behavior using histograms, scatter plots, and heatmaps to
identify trends and correlations between different product categories.

3. Modify:

- In the modify stage, data preprocessing techniques are applied to prepare the data for modeling.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

- Tasks include handling missing values, removing outliers, transforming variables, and creating new
features.

- Example: Cleaning the customer data by removing duplicate entries, filling in missing values in
demographic information, and transforming categorical variables into numerical representations.

4. Model:

- In the model stage, predictive models are built using various algorithms such as decision trees,
logistic regression, or neural networks.

- These models are trained on the prepared dataset to predict outcomes or classify data into
categories.

- Example: Building a predictive model to forecast customer churn based on demographic information,
purchase history, and engagement metrics.

5. Assess:

- In the assess stage, the performance of the models is evaluated using validation techniques and
evaluation metrics.

- Models are tested on unseen data to assess their accuracy, precision, recall, or other relevant
metrics.

- Example: Evaluating the performance of the churn prediction model using a holdout dataset and
metrics such as accuracy, precision, and recall to determine its effectiveness in identifying at-risk
customers.

By following the SEMMA approach, organizations can systematically analyze data, develop predictive
models, and derive actionable insights to drive decision-making and achieve business objectives.

Sampling:

Data sampling is the process of selecting a subset of data from a larger dataset to perform analysis or
build models. The goal of sampling is to obtain a representative sample that accurately reflects the
characteristics of the entire population while reducing computational complexity and processing time.

Concept: Data sampling involves randomly selecting a portion of data points from a population to make
inferences or draw conclusions about the entire population. The key is to ensure that the sample is
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

representative of the population, meaning it captures the same distribution and characteristics as the
original dataset.

Example: Consider a large dataset containing information about customer transactions in a retail store.
The dataset includes details such as customer demographics, purchase history, and product preferences.
Analyzing the entire dataset might be time-consuming and computationally intensive. Instead, you can
use data sampling to select a representative subset of customers for analysis.

Let's say you want to analyze customer purchasing behavior to identify trends and patterns. You could
employ random sampling by randomly selecting a certain percentage of customers from the dataset.
This sample should include customers from different demographics, geographic locations, and
purchasing habits to ensure representativeness.

For instance, if the original dataset contains 10,000 customer records, you might decide to sample 20%
of the data, resulting in a sample size of 2,000 customer records. These 2,000 records would then be
used for analysis, such as identifying popular products, analyzing spending patterns, or segmenting
customers based on their behavior.

By sampling the data, you can gain insights into customer behavior and preferences without having to
analyze the entire dataset. Additionally, sampling allows you to reduce computational resources and
processing time while still making informed decisions based on the characteristics of the population.

Data Pre-processing:

Data preprocessing is a crucial step in data mining and machine learning, involving the cleaning,
transformation, and preparation of raw data to make it suitable for analysis.

Concept: Data preprocessing involves several tasks to ensure that the dataset is clean, consistent, and
ready for analysis. These tasks may include handling missing values, dealing with outliers, scaling or
normalizing features, and encoding categorical variables.

Example: Consider a dataset containing information about housing prices, including features such as
square footage, number of bedrooms, location, and sale prices. Before building a predictive model to
estimate house prices, the dataset needs to undergo preprocessing.

1. Handling Missing Values:

 In the housing dataset, some entries may have missing values for certain features, such
as the number of bedrooms. One approach to handling missing values is to impute them
with the mean, median, or mode of the respective feature.

2. Dealing with Outliers:


GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

 Outliers in the dataset, such as unusually high or low sale prices, can skew the results of
the analysis. Outliers can be identified using statistical methods like Z-score or by visual
inspection using box plots. Depending on the situation, outliers can be removed or
transformed.

3. Scaling or Normalizing Features:

 Features in the dataset may have different scales, making some features dominate
others during analysis. Scaling or normalizing features ensures that all features
contribute equally to the analysis. For example, the square footage feature may be
scaled to have a mean of 0 and a standard deviation of 1.

4. Encoding Categorical Variables:

 Categorical variables, such as location or neighborhood names, need to be encoded into


numerical representations for analysis. This can be done using techniques like one-hot
encoding or label encoding. For instance, the location feature with categories like
"Suburban," "Urban," and "Rural" could be encoded as binary variables (e.g., 0 or 1).

5. Feature Selection/Extraction:

 Not all features in the dataset may be relevant for predicting house prices. Feature
selection techniques, such as correlation analysis or feature importance ranking, can
help identify the most important features. Additionally, feature extraction techniques
like Principal Component Analysis (PCA) can be used to reduce the dimensionality of the
dataset while preserving its essential information.

By preprocessing the housing dataset, it becomes cleaner, more consistent, and suitable for building
predictive models to estimate house prices accurately. Data preprocessing ensures that the dataset's
quality is improved and that the subsequent analysis or modeling tasks produce reliable results.

Data Cleaning:
Data cleaning is the process of identifying and correcting errors, inconsistencies, and missing values in a
dataset to ensure its accuracy and reliability.

Concept: Data cleaning involves various tasks such as handling missing values, removing duplicates,
correcting errors, and standardizing formats to improve the quality of the dataset. The goal is to prepare
the data for analysis or modeling by ensuring that it is accurate, complete, and consistent.

Example: Consider a dataset containing information about customer orders from an e-commerce
website. The dataset includes details such as customer IDs, order IDs, product names, quantities, prices,
and timestamps. Before analyzing this dataset, it needs to undergo data cleaning to address any issues.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

1. Handling Missing Values:

 The dataset may contain missing values in certain fields, such as the product name or
quantity. These missing values need to be identified and addressed. For example,
missing product names could be imputed with the most common product in the dataset,
while missing quantities could be imputed with the median quantity.

2. Removing Duplicates:

 Duplicate entries, where identical records appear more than once in the dataset, need
to be identified and removed. Duplicate orders or customer IDs could skew the analysis
results. Removing duplicates ensures that each record in the dataset is unique.

3. Correcting Errors:

 Errors in the dataset, such as typos, inconsistencies, or incorrect data entries, need to be
corrected. For instance, if there are inconsistencies in the product names (e.g., "iPhone
12" vs. "iphone 12"), they need to be standardized to ensure consistency.

4. Standardizing Formats:

 Date and time formats, currency symbols, and other formats in the dataset may vary.
Standardizing these formats ensures consistency across the dataset. For example,
timestamps could be converted to a uniform format (e.g., YYYY-MM-DD HH:MM:SS).

5. Validation:

 After cleaning the data, it's important to validate the cleaned dataset to ensure that the
cleaning process was successful and that the data is now ready for analysis. This may
involve spot-checking a sample of records or running validation checks on key fields.

By cleaning the dataset, inconsistencies, errors, and missing values are addressed, resulting in a dataset
that is accurate, complete, and reliable. This cleaned dataset can then be used for further analysis,
modeling, or reporting, leading to more accurate and actionable insights.

Data Partitioning:

Data partitioning, also known as data splitting, is the process of dividing a dataset into two or more
subsets for different purposes, such as training, validation, and testing. Each subset serves a specific role
in the machine learning workflow.

Concept: Data partitioning involves splitting a dataset into multiple subsets to facilitate model
development, evaluation, and testing. The most common partitions are:
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

1. Training Set: The training set is used to train the machine learning model. It contains a majority
of the data and is used to learn the patterns and relationships between features and labels.

2. Validation Set: The validation set is used to tune hyperparameters and assess the performance
of the model during training. It helps prevent overfitting by providing an independent dataset
for model evaluation.

3. Test Set: The test set is used to evaluate the final performance of the trained model. It serves as
an unbiased measure of the model's generalization ability on unseen data.

Example: Consider a dataset containing information about housing prices, including features such as
square footage, number of bedrooms, location, and sale prices. Before building a predictive model to
estimate house prices, the dataset needs to be partitioned into training, validation, and test sets.

1. Training Set:

 70-80% of the dataset is randomly selected and used as the training set.

 This subset contains the majority of the data and is used to train the machine learning
model.

 Example: If the original dataset contains 1,000 records, 700-800 records are randomly
sampled and used for training the model.

2. Validation Set:

 10-15% of the dataset is randomly selected and used as the validation set.

 This subset is used to tune hyperparameters and evaluate the model's performance
during training.

 Example: If the original dataset contains 1,000 records, 100-150 records are randomly
sampled and used for validation.

3. Test Set:

 The remaining portion of the dataset (10-20%) is used as the test set.

 This subset is used to evaluate the final performance of the trained model.

 Example: If the original dataset contains 1,000 records, 100-200 records are reserved for
testing the model's performance.

By partitioning the dataset into training, validation, and test sets, we ensure that the model is trained on
one subset, fine-tuned on another, and evaluated on a completely independent subset. This helps assess
the model's performance accurately and provides insights into its generalization ability on unseen data.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

Building a Data Model:

Building a data model involves creating a mathematical representation of the relationships and patterns
in a dataset, typically through the use of machine learning algorithms.

Building a data model involves several steps, including selecting an appropriate algorithm, preparing the
data, training the model, tuning hyperparameters, and evaluating its performance.

1. Selecting an Algorithm:

 Choose a machine learning algorithm suitable for the task at hand, such as regression,
classification, clustering, or anomaly detection, based on the nature of the data and the
desired outcome.

2. Data Preparation:

 Preprocess the data by cleaning, transforming, and encoding features as necessary to


ensure they are in a format suitable for the chosen algorithm. This may involve tasks
such as handling missing values, scaling features, and encoding categorical variables.

3. Training the Model:

 Split the dataset into training and validation sets.

 Train the selected algorithm on the training data, where the model learns the patterns
and relationships between features and labels. The algorithm adjusts its parameters
iteratively to minimize the error between predicted and actual outcomes.

4. Hyperparameter Tuning:

 Fine-tune the model's hyperparameters to optimize its performance. Hyperparameters


are parameters that govern the learning process, such as the learning rate,
regularization strength, or tree depth.

 This is typically done using techniques like grid search, random search, or Bayesian
optimization to search the hyperparameter space and find the optimal configuration.

5. Model Evaluation:

 Evaluate the performance of the trained model on the validation set using appropriate
metrics for the task, such as accuracy, precision, recall, F1-score, or mean squared error.

 Compare the model's performance against baseline models or other algorithms to


assess its effectiveness and generalization ability.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

6. Model Deployment:

 Once satisfied with the model's performance, deploy it into production to make
predictions on new, unseen data.

 Integrate the model into existing systems or applications for real-time decision-making
or automate processes based on the model's predictions.

Example: Suppose you're building a predictive model to forecast sales revenue for an e-commerce
company based on historical sales data, website traffic, marketing spend, and other relevant features.

1. Selecting an Algorithm:

 Choose a regression algorithm, such as linear regression or decision trees, suitable for
predicting numerical values.

2. Data Preparation:

 Clean the sales data, handle missing values, and encode categorical variables like
product categories.

 Scale numerical features like website traffic and marketing spend to ensure they have
similar magnitudes.

3. Training the Model:

 Split the dataset into training and validation sets.

 Train the regression model on the training data to learn the relationships between
features and sales revenue.

4. Hyperparameter Tuning:

 Tune hyperparameters like regularization strength or tree depth to optimize the model's
performance.

5. Model Evaluation:

 Evaluate the model's performance on the validation set using metrics like mean squared
error or R-squared to assess its accuracy and predictive power.

6. Model Deployment:

 Once validated, deploy the trained regression model into production to predict sales
revenue for future periods based on new data inputs.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

By following these steps, you can build a data model that accurately captures the patterns and
relationships in the data, allowing you to make informed decisions and generate valuable insights.

Statistical Models for Predictive Analysis:

Statistical models for predictive analytics are mathematical representations of relationships and
patterns in data, used to make predictions or forecasts about future outcomes based on historical data.
Here's a brief description of some common statistical models along with examples:

1. Linear Regression:

 Linear regression models the relationship between a dependent variable and one or
more independent variables by fitting a linear equation to the observed data.

 Example: Predicting house prices based on features like square footage, number of
bedrooms, and location. The linear regression model estimates the relationship
between these features and the house price.

2. Logistic Regression:

 Logistic regression is used for binary classification tasks, where the dependent variable
is categorical and has only two outcomes.

 Example: Predicting whether a customer will churn (leave) a subscription service based
on demographic information and usage patterns. The logistic regression model
estimates the probability of churn based on these features.

3. Decision Trees:

 Decision trees partition the feature space into regions based on the values of input
variables, leading to a tree-like structure of decision rules.

 Example: Predicting whether a loan applicant will default based on features like credit
score, income, and debt-to-income ratio. The decision tree model creates a set of rules
to classify loan applicants as high-risk or low-risk.

4. Random Forests:

 Random forests are an ensemble learning method that combines multiple decision trees
to improve predictive performance and reduce overfitting.
GLOBAL ACADEMY OF TE CHNOLOGY,
BENGALURU
( An A ut on omous In sti t ute , affi l i ate d to VT U, Be l agavi )

 Example: Predicting customer sentiment (positive or negative) based on reviews and


social media posts. The random forest model aggregates predictions from individual
decision trees to make a final prediction about sentiment.

5. Support Vector Machines (SVM):

 SVM is a supervised learning algorithm that finds the optimal hyperplane in a high-
dimensional feature space to separate data points into different classes.

 Example: Predicting whether an email is spam or not based on features extracted from
the email content. The SVM model learns a decision boundary to classify emails into
spam or non-spam categories.

6. Time Series Models:

 Time series models are used to analyze and forecast data points collected over time,
capturing temporal dependencies and trends.

 Example: Forecasting stock prices based on historical price data. Time series models like
ARIMA (Autoregressive Integrated Moving Average) or LSTM (Long Short-Term Memory)
can be used to predict future stock prices based on past trends.

These are just a few examples of statistical models commonly used in predictive analytics. Each model
has its strengths and limitations, and the choice of model depends on the nature of the data, the specific
prediction task, and the desired accuracy of the predictions.

You might also like