Unit Ii
Unit Ii
Unit Ii
UNIT-II
Analytical methods and models: Descriptive analytics-overview of its tools and techniques, role in business
analytics process and its importance in business decision making; Predictive analytics-nature and type of
modelling, basics of data mining and machine learning environment, role in business analytics process and its
importance in strategic decision making; Prescriptive analytics: basics of its tools and modelling, role in business
analytics process.
Descriptive analytics provides the necessary context and understanding for your organization
to then diagnose, predict, and prescribe actions based on data-driven insights, forming a
comprehensive analytical framework.
How It Works
Descriptive analytics works by analyzing and summarizing historical data to provide insights
into past events, patterns, and trends. This is why it’s more similar to reporting vs analytics as
most people think of it. The process typically involves several steps:
Collect Data: First, you collect relevant data from various operational sources, such as
databases, spreadsheets, or data warehouses. This data can include structured data such as
numerical values, categorical variables, as well as unstructured data like text or images.
Clean & Prep Data: You then clean and process this data to ensure accuracy and consistency.
This step involves handling missing values, removing duplicates, and transforming the data
into a suitable format for analysis. This step can happen before or after the data is loaded into
a data repository.
Explore Data: Apply exploratory data analysis techniques to gain a better understanding of
the dataset. This includes examining summary statistics, distributions, and visualizations to
identify patterns, outliers, and relationships within the data.
Summarize Data: Calculate descriptive statistics to summarize the dataset, providing key
metrics such as mean, median, mode, standard deviation, and percentiles. These statistics offer
a concise overview of the dataset's central tendencies and dispersion.
Visualize Data: Create visual representations, such as charts, graphs, and dashboards, to
present the data in a visually appealing and easily understandable format. Visualizations help
identify trends, patterns, and anomalies more intuitively.
Interpret Data: Interpret the summarized data and visualizations to extract meaningful
insights and observations about the past events and trends. This analysis enables stakeholders
to understand the data's implications and make informed decisions based on the findings.
Advantages:
• Provides a historical perspective: Descriptive analytics allows businesses to gain
insights into past events, patterns, and trends, providing a valuable historical context
for decision-making.
• Easy to understand and interpret: Summarized and visualized data makes it
accessible to a wide range of users, enabling easy understanding and interpretation of
the information.
3
Introduction
As businesses continue to embrace digital transformation and data-driven
decision-making, predictive analytics has emerged as a game-changer. AI-driven predictive
models allow you to harness the power of historical data and statistical algorithms to
make predictions about the future with unparalleled accuracy in mere seconds.
Whether you’re looking to optimize your marketing campaigns, reduce operational costs,
or improve customer experience, predictive analytics can help you achieve your targets. In
this blog, we’ll explore the various types of predictive analytics models, how they work,
their applications, their benefits, and their limitations. By the end of this blog, you’ll have a
comprehensive understanding of how predictive analytics models can elevate your
business and drive better outcomes. Let’s dive in.
What is Predictive Analytics?
Predictive analytics is a subfield of data science that analyses historical data and makes
predictions about future events or outcomes using statistical algorithms, machine
learning, and other techniques. This method assists organizations in identifying patterns,
detecting anomalies, and forecasting future trends.
To read more on predictive analytics, click here for our detailed blog.
5
Learn about how you can predict and prevent churn with
Voxco Intelligence.
Book a Demo
What are Predictive Analytics Models?
Predictive analytics models are statistical or machine learning algorithms that use
historical data to make predictions about future events or behaviors. These models
analyze large amounts of data to identify patterns and trends, which they then use to
forecast future outcomes.
Predictive analytics model development typically entails several steps, including data
preparation, model selection, feature engineering, and model training and testing. The
goal is to create a model that predicts future outcomes accurately based on a set of input
variables or features.
Finance, healthcare, marketing, and retail are just a few of the industries that can benefit
from predictive analytics modeling. Predictive modeling in business analytics can help in
identifying potential risks and opportunities, optimizing operations, and making more
informed decisions based on data. Read our article on the use of predictive analytics in
business to get more insights.
How Do Predictive Analytics Models Work?
Predictive analytics models identify patterns and trends in historical data that can be used
to predict future outcomes. Several steps are usually involved in the process:
Gathering relevant data from multiple sources is the first step in developing a predictive
analytics model.
Once the data has been collected, it must be cleaned and transformed into an
analysis-ready format.
In this step, the most relevant variables or features are selected from the dataset to be
used in the model.
There are various types of predictive analytics models available, including regression,
decision trees, and neural networks. Based on the nature of the problem and the
available data, the appropriate model is chosen.
The model is trained on historical data in this step to learn patterns and relationships that
can be used to make predictions.
The model is evaluated after it has been trained to ensure that it can make accurate
predictions on new, unseen data.
Finally, the model is used to make predictions based on new data and provide business
insights.
1. Regression models: These models use one or more input variables to predict a
continuous numeric value, such as sales revenue or customer lifetime value.
2. Classification models: these are used to categorize data into one of two or more
groups based on input variables. Fraud detection, customer segmentation, and
spam filtering are a few examples.
3. Time series models: These models are used to predict future values based on
historical trends and patterns in time-series data, such as stock prices, weather
patterns, or website traffic.
4. Clustering models: These models are used to categorize data points based on
similarities in their characteristics or behaviors. Clustering models are often used
for customer segmentation and market basket analysis.
5. Neural network models: These models are used to identify complex patterns in
data and are based on the structure and function of the human brain. Image and
speech recognition, natural language processing, and predictive maintenance are all
common applications for neural network models.
These are a few predictive analytics models examples. There are numerous other models,
each with its own set of strengths and weaknesses, that can be used to solve various
types of problems.
What are the 3 Levels of Predictive Modeling?
Predictive modeling is divided into three levels, each with a different level of complexity
and accuracy.
This is the most basic level of predictive modeling, involving the use of historical data to
identify patterns and trends. The goal of descriptive modeling is to comprehend what has
occurred in the past and to provide insights into what is likely to occur in the future based
on that data.
This level of predictive modeling entails making predictions about future events or
outcomes using statistical algorithms and machine learning techniques. Predictive models
are trained on historical data to identify patterns and relationships between various data
points. Predictive modeling’s goal is to make accurate predictions about future events
based on historical data.
The most advanced level of predictive modeling, in which predictive models are used to
make recommendations or decisions. Prescriptive models consider a variety of variables
and factors before recommending the best course of action to achieve a specific goal or
outcome. Prescriptive modeling seeks to improve decision-making processes and
maximize business outcomes.
Applications of Predictive Analytics Models
Businesses can benefit from predictive analytics models in a variety of ways, including:
Although predictive models offer a wide range of benefits, they also a couple limitations
that must be considered. Here are some of the key limitations of predictive analytics
models.
algorithm for the problem being addressed and ensure that it is regularly
updated and refined as new data becomes available.
Predictive analytics models are powerful tools that can help businesses gain insights into
future trends and make informed decisions. Businesses can gain a competitive advantage
in their industry by leveraging historical data and advanced algorithms. Companies in the
financial services, healthcare, and retail industries benefit greatly from the use of
predictive analytics software, predictive analytics tools are a type of tool that is useful to
businesses of all sizes.
The AI capabilities of Voxco Intelligence enable you to build predictive models that deliver
predictive insights into customer behavior, make future forecasts, and make data-driven
decisions to improve customer experience.
The best model for predictive analytics depends on the nature of the problem and the
available data. There are various types of predictive models, each with its own set of
advantages and disadvantages, and selecting the right one is dependent on the specific
needs of the business. It is critical to select a model that is appropriate for the problem
and the available data, as well as to continuously evaluate and refine the model in order
to improve its accuracy and performance.
Data Mining:
The origins of Data Mining
10
✓ Data mining is a discipline with a long history. It starts with the early Data Mining
methods Bayes’ Theorem (1700`s) and Regression analysis (1800`s)
which were mostly identifying patterns in data.
✓ Data mining is the process of analyzing large data sets (Big Data) from different
perspectives and uncovering correlations and patterns to summarize them into useful
information.
✓ In the late 80`s Data Mining term began to be known and used within the research
community by statisticians, data analysts, and the management information systems
(MIS) communities.
✓ By the early 1990`s, data mining was recognized as a sub-process or a step within a
larger process called Knowledge Discovery in Databases (KDD) .The most commonly
used definition of KDD is “The nontrivial process of identifying
valid, novel, potentially useful, and ultimately understandable patterns in data”
(Fayyad, 1996).
1. Understanding of the application and identifying the goal of the KDD process
4. Matching the goals of the KDD process (step 1) to a particular data-mining method.
Data mining functionalities are to perceive the various forms of patterns to be identified in
data mining activities. To define the type of patterns to be discovered in
data mining activities, data mining features are used. Data mining has a wide application
for forecasting and characterizing data in big data.
Data mining tasks are majorly categorized into two categories: descriptive and predictive.
2) Prediction
To detect the inaccessible data, it uses regression analysis and detects the missing numeric
values in the data. If the class mark is absent, so classification is used to render the prediction.
Due to its relevance in business intelligence, the prediction is common. If the class mark is
absent, so the prediction is performed using classification. There are two
methods of predicting data. Due to its relevance in business intelligence, a prediction is
common. The prediction of the class mark using the previously developed class model and the
prediction of incomplete or incomplete data using prediction analysis are two ways of
predicting data.
14
15
3) Classification
4) Association Analysis
The link between the data and the rules that bind them is discovered. And two
or more data attributes are associated. It associates qualities that are
transacted together regularly. They work out what are called the rules of
partnerships that are commonly used in the study of stock baskets. To link
the attributes, there are two elements. One is the trust that suggests the
possibility of both associated together, and another helps, which informs of
associations' past occurrence.
5) Outlier Analysis
Data components that cannot be clustered into a given class or cluster are
outliers. They are often referred to as anomalies or surprises and are also very
important to remember.
Although in some contexts, outliers can be called noise and discarded, they
can disclose useful information in other areas, and hence can be very
important and beneficial for their study.
6) Cluster Analysis
We may uncover patterns and shifts in actions over time, with such distinct
analysis, we can find features such as time-series results, periodicity,
and similarities in patterns. Many technologies from space science
to retail marketing can be found holistically in data processing
and features.
❖ Machine Learning
3. Semi-supervised learning
Duda and Hart in 1973. In 1981 a report was given on using teaching
strategies so that a neural network learns to recognize 40 characters
(26 letters, 10 digits, and 4 special symbols) from a computer terminal.
TensorFlow
Theano Theano is wonderfully folded over Keras, an abnormal state neural systems
library, that runs nearly in parallel with the Theano library. Keras’ fundamental
favorable position is that it is a moderate Python library for profound discovering
that can keep running over Theano or TensorFlow.
Sci-Kit Learn
This framework involves a lot of calculations for regular AI and data mining
assignments, including bunching, relapse, and order.
Caffe
H20
It is a service that makes it easy for developers of all skill levels to use
machine learning technology. It connects to data stored in Amazon S3,
Redshift, or RDS, and can run binary classification, multiclass categorization,
or regression on the data to build a model.
Torch
The goal of Torch is to have maximum flexibility and speed in building your
scientific algorithms along with an extremely simple process.
Azure ML Studio
This Framework allows Microsoft Azure users to create and train models, then
turn them into APIs that can be consumed by other services. Also, you can
connect your own Azure storage to the service for larger models.
To use the Azure ML Studio, you don’t even need an account to try out the
service. You can log in anonymously and use Azure ML Studio for up to eight
hours.
Spark ML Lib
This is Apache Spark’s machine learning library. The goal of this framework
is to make practical machine learning scalable and easy.
The details of your process will vary depending on your specific use case
and type of data but below is a high-level overview to get you started.
3) Develop your model. Now you’re ready to build, train, evaluate and
deploy your prescriptive model. You can hire a data scientist to code one
from scratch or you can leverage an AutoML tool to develop a custom ML
model yourself as a citizen data scientist. Either way, this algorithm-based
model will need to ingest a mix of structured data, unstructured data, and
defined business rules. Analytic techniques used in your model may include
simulation, graph analysis, heuristics, optimization, and game theory. You’ll
have to tweak your model in iterations to get it right and you’ll definitely
want to test your model multiple times using new data to see if the
recommendations generated meet what you would expect.
4) Deploy your model. Once you’re confident in its performance, you can
make your prescriptive model available for use. This may be a one-time
project or as part of an on-going production process. For a one-time
project, an asynchronous batch recommendation is probably most
appropriate. If your model will be integral to a larger process in which
other applications depend on fast predictions, a synchronous, real-time
deployment is best. Your model should automatically adjust as new data is
added over time. This will improve the accuracy of the recommendations.
5) Take action. Now you should review the recommendation, decide if it
makes sense to you, and then take appropriate actions. Some situations
require human intuition and judgment and in these cases, prescriptive
analytics should be viewed as decision support rather than decision
automation. Conversely, if your prescriptive model is integrated to a larger
process, the downstream actions may happen automatically. Application
26
automation triggering action is one of the top 10 BI and data trends this
year.
Prescriptive Analytics Challenges
• Certain situations require human intuition and
judgment. Sometimes you’ll need to bring additional context such
as knowing external market conditions or competitor activities
which may not be included in the algorithm. In these cases,
prescriptive analytics should be viewed as decision support rather
than decision automation.
• Invalid inputs result in invalid outputs. Even the best models will
produce poor recommendations if the data and/or assumptions
built into your model are not accurate.
• Avoid overkill. Not every organization, campaign, or situation
requires the power of prescriptive analysis. In some cases, the
effort of setting it up may not be worth it.
• Models take time to refine. Training and evaluating your model is
a critical step in getting prescription analytics right. Even with
AutoML tools which get you started with a default model, you
should adjust parameters to better fit your specific use case and
test evaluate your model using new data to see if the
recommendations generated meet what you’d expect.