Chapter 3545
Chapter 3545
ANALYTICS
As per New CBCS Syllabus for BBA, 3rd Year, 6th Semester for
All the Universities in Telangana State w.e.f. 2018-19
V.Karuna Sree
MBA, M.Com, PGD(Maths) (PhD)
Academic Head,
Ethemes College of Commerce & Business Management,
Panjagutta, Hyderabad.
Published by : Mrs. Meena Pandey for Himalaya Publishing House Pvt. Ltd.,
“Ramdoot”, Dr. Bhalerao Marg, Girgaon, Mumbai - 400 004.
Phone: 022-23860170, 23863863; Fax: 022-23877178
E-mail: [email protected]; Website: www.himpub.com
Branch Offices :
New Delhi : “Pooja Apartments”, 4-B, Murari Lal Street, Ansari Road, Darya Ganj, New Delhi - 110 002.
Phone: 011-23270392, 23278631; Fax: 011-23256286
Nagpur : Kundanlal Chandak Industrial Estate, Ghat Road, Nagpur - 440 018.
Phone: 0712-2738731, 3296733; Telefax: 0712-2721216
Bengaluru : Plot No. 91-33, 2nd Main Road, Seshadripuram, Behind Nataraja Theatre,
Bengaluru - 560 020. Phone: 080-41138821; Mobile: 09379847017, 09379847005
Chennai : New No. 48/2, Old No. 28/2, Ground Floor, Sarangapani Street, T. Nagar,
Chennai - 600 012. Mobile: 09380460419
Pune : “Laksha” Apartment, First Floor, No. 527, Mehunpura, Shaniwarpeth (Near Prabhat Theatre),
Pune - 411 030. Phone: 020-24496323, 24496333; Mobile: 09370579333
Lucknow : House No. 731, Shekhupura Colony, Near B.D. Convent School, Aliganj,
Lucknow - 226 022. Phone: 0522-4012353; Mobile: 09307501549
Ahmedabad : 114, “SHAIL”, 1st Floor, Opp. Madhu Sudan House, C.G. Road, Navrang Pura,
Ahmedabad - 380 009. Phone: 079-26560126; Mobile: 09377088847
Ernakulam : 39/176 (New No. 60/251), 1st Floor, Karikkamuri Road, Ernakulam, Kochi - 682 011.
Phone: 0484-2378012, 2378016; Mobile: 09387122121
Bhubaneswar : Plot No. 214/1342, Budheswari Colony, Behind Durga Mandap, Bhubaneswar - 751 006.
Phone: 0674-2575129; Mobile: 09338746007
Kolkata : 108/4, Beliaghata Main Road, Near ID Hospital, Opp. SBI Bank, Kolkata - 700 010.
Phone: 033-32449649; Mobile: 07439040301
DTP by : Nilima
Printed at : M/s. Aditya Offset Process (I) Pvt. Ltd., Hyderabad. On behalf of HPH.
Preface
The objective of this book is to provide an understanding of basic concepts of Business Analytics
like Descriptive, Predictive and Prescriptive Analytics, Data Mining Techniques and an overview of
Programming using R.
The related matter had been written in a simple and lucid style, easily understandable language
even for the below-average students with sufficient support from real business information.
This book covers five units:
Unit-I: Business Analytics, Categories of Business Analytical Methods and Models, Business
Analytics in Practice, Big Data – Overview of Using Data, Types of Data.
Unit-II: Description Statistics (Central Tendency, Variability), Data Visualization – Definition,
Visualization Techniques – Tables, Cross Tabulations, Charts, Data Dashboards using MS-Excel or
SPSS.
Unit-III: Trend Lines, Regression Analysis – Linear and Multiple, Forecasting Techniques, Data
Mining – Definition, Approaches in Data Mining – Data Exploration and Reduction, Classification,
Association, and Cause and Effect Modeling.
Unit-IV: Linear Optimization, Non-linear Programming, Integer Optimization, Cutting Plane
Algorithm and Other Methods, Decision Analysis – Risk and Uncertainty Methods.
Unit-V: Programming using R, R Environment, R Packages, Reading and Writing Data in R, R
Functions, Control Statements, Frames and Subsets, Managing and Manipulating Data in R.
Also, we had given MCQs.
We consider this book is useful for understanding purposes to students and professionals. This
book lays down the framework defining Descriptive, Predictive and Prescriptive Analytics.
And our wholehearted gratitude to FR. Rev. Allam. Arogya Reddy, Principal, St. Mary’s
Centenary Degree colLege and Shri. K.Raghuveer Principal, IIMC, Hyderabad and lso to my dear
colleagues and to our family members.
We offer our gratitude to Himalaya Publishing House Pvt. Ltd., who is a leader in Commerce and
Management Publications. Our sincere regards to Shri Niraj Pandey (Director), Vijay Pandey (General
Manager – Marketing) and especially Mr. G. Anil Kumar (Sales Manager), Hyderabad for his sincere
efforts and interest shown and for the best efforts and patience taken for bringing out this book it time.
Any query or any suggestions regarding improvement or errors, if any, will be gratefully
acknowledged and shall be incorporated in our consecutive editions
You can contact on [email protected].
January 2019
Hyderabad Authors
Syllabus
Objective: The objective of the course is to provide an understanding of basic concepts of
Business Analytics like Descriptive, Predictive and Prescriptive Analytics and an overview of
Programming using R.
2 Descriptive Analytics 23 – 89
Chapter Outline
1.1 Definition of Business Analytics
1.2 Categories of Business Analytical Methods and Models
1.3 Business Analytics in Practice
1.4 Big Data – Overview of Using Data
1.5 Types of Data
1.6 Questions
What’s happening,
why is it happening, Predict opportunities in which
2. Predictive analytic
and what will happen? analysis the firm can take advantage.
The second area of business analytics involves deeper statistical analysis. This may mean doing
predictive analytics by applying statistical algorithms to historical data to make a prediction about
future performance of a product, service or website design change. Or, it could mean using other
advanced analytics techniques like cluster analysis, to group customers based on similarities across
several data points. This can be helpful in targeted marketing campaigns, for example.
True data science involves more custom coding and more open-ended questions. Data scientists
generally do not set out to solve a specific question, as most business analysts do. Rather, they will
explore data using advanced statistical methods and allow the features in the data to guide their
analysis.
business systems, data cleansing and integration into a single repository, such as a data warehouse or
data mart. The analysis is typically performed against a smaller sample set of data.
Analytics tools range from spreadsheets with statistical functions to complex data mining and
predictive modeling applications. As patterns and relationships in the data are uncovered, new
questions are asked, and the analytical process iterates until the business goal is met.
Deployment of predictive models involves scoring data records – typically in a database – and
using the scores to optimize real-time decisions within applications and business processes. BA also
supports tactical decision-making in response to unforeseen events. And, in many cases, the decision-
making is automated to support real-time responses.
Discussion Questions
1. What is the difference between analytics and business analytics?
2. What is the difference between business analytics and business intelligence?
3. Why are the steps in the business analytics process sequential?
4. How is the business analytics process similar to the organization decision-making process?
Descriptive Prescriptive
Diagnostic Predictive
recommends
explains what explains why it forecasts what
an action based on
happened. happened. might happen.
the forecast
1. Descriptive Analytics
This can be termed as the simplest form of analytics. The mighty size of big data is beyond
human comprehension and the first stage, hence involves crunching the data into understandable
chunks. The purpose of this analytics type is just to summarize the findings and understand what is
going on.
Among some frequently used terms, what people call as advanced analytics or business
intelligence is basically usage of descriptive statistics (arithmetic operations, mean, median, max,
percentage, etc.) on existing data. It is said that 80% of business analytics mainly involves descriptions
based on aggregations of past performance. It is an important step to make raw data understandable to
investors, shareholders and managers. This way, it gets easy to identify and address the areas of
strengths and weaknesses such that it can help in strategizing.
The two main techniques involved are data aggregation and data mining stating that this method
is purely used for understanding the underlying behavior and not to make any estimations. By mining
historical data, companies can analyze the consumer behaviors and engagements with their businesses
that could be helpful in targeted marketing, service improvement, etc. The tools used in this phase are
MS Excel, MATLAB, SPSS, STATA, etc.
2. Diagnostic Analytics
Diagnostic Analytics is used to determine why something happened in the past. It is characterized
by techniques such as drill-down, data discovery, data mining and correlations. Diagnostic analytics
takes a deeper look at data to understand the root causes of the events. It is helpful in determining what
factors and events contributed to the outcome. It mostly uses probabilities, likelihoods and the
distribution of outcomes for the analysis.
In a time series data of sales, diagnostic analytics would help you understand why the sales have
decreased or increased for a specific year or so. However, this type of analytics has a limited ability to
6 Business Analytics
give actionable insights. It just provides an understanding of causal relationships and sequences while
looking backward.
A few techniques that uses diagnostic analytics include attribute importance, principle
components analysis, sensitivity analysis and conjoint analysis. Training algorithms for classification
and regression also fall in this type of analytics.
3. Predictive Analytics
Predictive Analytics is used to predict future outcomes. However, it is important to note that
it cannot predict if an event will occur in the future; it merely forecasts what the probabilities of the
occurrence of the event are. A predictive model builds on the preliminary descriptive analytics stage to
derive the possibility of the outcomes.
The essence of predictive analytics is to devise models such that the existing data is understood to
extrapolate the future occurrence or simply, predict the future data. One of the common applications of
predictive analytics is found in sentiment analysis where all the opinions posted on social media are
collected and analyzed (existing text data) to predict the person’s sentiment on a particular subject as
being positive, negative or neutral (future prediction).
Hence, predictive analytics includes building and validation of models that provide accurate
predictions. Predictive analytics relies on machine learning algorithms like random forests, SVM, etc.
and statistics for learning and testing the data. Usually, companies need trained data scientists and
machine learning experts for building these models. The most popular tools for predictive analytics
include Python, R, Rapid Miner, etc.
The prediction of future data relies on the existing data as it cannot be obtained otherwise. If the
model is properly tuned, it can be used to support complex forecasts in sales and marketing. It goes
a step ahead of the standard BI in giving accurate predictions.
4. Prescriptive Analytics
The basis of this analytics is predictive analytics, but it goes beyond the three mentioned above to
suggest the future solutions. It can suggest all favorable outcomes according to a specified course of
action and also suggest various course of actions to get to a particular outcome. Hence, it uses a strong
feedback system that constantly learns and updates the relationship between the action and the
outcome.
The computations include optimization of some functions that are related to the desired outcome.
For example, while calling for a cab online, the application uses GPS to connect you to the correct
driver from among a number of drivers found nearby. Hence, it optimizes the distance for faster arrival
time. Recommendation engines also use prescriptive analytics.
The other approach includes simulation where all the key performance areas are combined to
design the correct solutions. It makes sure whether the key performance metrics are included in the
solution. The optimization model will further work on the impact of the previously made forecasts.
Because of its power to suggest favorable solutions, prescriptive analytics is the final frontier of
advanced analytics or data science, in today’s term.
Analytical Models
An analytical model is simply a mathematical equation that describes relationships among
variables in a historical data set. The equation either estimates or classifies data values. In essence,
a model draws a “line” through a set of data points that can be used to predict outcomes. For example,
a linear regression draws a straight line through data points on a scatterplot that shows the impact of
advertising spend on sales for various ad campaigns. The model’s formula—in this case, “Sales =
17.813 + (.0897 * advertising spend)”—enables executives to accurately estimate sales if they spend a
specific amount on advertising (See Figure 1.)
Estimation Model (Linear Regression)
Sales
Sales = 17.813 + (.0897 * Advertising Spend)
Advertising
Figure 1
Algorithms that create analytical models (or equations) come in all shapes and sizes.
Classification algorithms such as neural networks, decision trees, clustering and logistic regression use
a variety of techniques to create formulas that segregate data values into groups. Online retailers often
use these algorithms to create target market segments or determine which products to recommend to
buyers based on their past and current purchases (See Figure 2).
Classification of Algorithms
Figure 2
8 Business Analytics
6. Model management 9%
Other 8%
Figure 3
Introduction to Business Analytics 9
From Wayne Eckerson, “Predictive Analytics: Extending the Value of Your Data Warehousing
Investment,” 2007. Based on 166 respondents, who have a predictive modeling practice.
Project Definition. Although defining an analytical project does not take as long as some of the
other steps, it is the most critical task in the process. Modelers that do not know explicitly what they
are trying to accomplish will not be able to create useful analytical models. Thus, before they start,
good analytical modelers spend a lot of time defining objectives, impact and scope.
Project Objectives. Project Objectives consist of the assumptions or hypotheses that a model
will evaluate. Often, it helps to brainstorm hypotheses and then prioritize them based on business
requirements. Project impact defines the model output (e.g., a report, a chart, or scoring program), how
the business will use that output (e.g., embedded in a daily sales report or operational application or
used in strategic planning), and the projected return on investment. Project scope defines who, what,
where, when, why and how of the project, including timelines and staff assignments.
For example, a project objective might be: “Reduce the amount of false positives when scanning
credit card transactions for fraud.” While the output might be: “A computer model capable of running
on a server and measuring 7,000 transaction per minute, scoring each with probability and confidence,
and routing transactions above a certain threshold to an operator for manual intervention.”
Data Exploration. Data exploration or data discovery involves sifting through various sources of
data to find the data sets that best fit the project. During this phase, the analytical modeler will
document each potential data set with the following items:
Access methods: Source systems, data interfaces, machine formats (e.g., ASCII or
EBCDIC), access rights and data availability.
Data characteristics: Field names, field lengths, content, format, granularity and statistics
(e.g., counts, mean, mode, median, and min/max values).
Business rules: Referential integrity rules, defaults and other business rules.
Data pollution: Data entry errors, misused fields and bogus data.
Data completeness: Empty or missing values and sparsely.
Data consistency: Labels and definitions.
Typically, an analytical modeler will compile all this information into a document and use it to
help prioritize which data sets to use for which variables (See Figure 4). A data warehouse with well
documented meta data can greatly accelerate the data exploration phase because it also maintains
much of this information. However, analytical modelers often want to explore external data and other
data sets that do not exist in the data warehouse and must compile this information manually.
Data Profile Document
Reverse Pivoting
Bank Transactions by Account Summarized Customer Activity
Period 1 - #
Period 1 - $
Period 2 - #
Period 2 - $
Period 3 - #
Period 3 - $
Period 4 - #
Period 4 - $
Product
Balance
Branch
Acct #
Acct #
Date
6651 3/17 $950 ELA ATM 6651 2 $65 3 $115 1 $100 5 $250
Figure 4
Introduction to Business Analytics 11
To model a banking “customer” not bank transactions, analytical modelers use a technique
called reverse pivoting to summarize banking transactions to show customer activity by period.
Analytical Modeling. Analytical modeling is as much art as science. Much of the craft involves
knowing what data sets and variables to select and how to format and transform the data for specific
data models. Often, a modeler will start with 100+ variables and then, through data transformation and
experimentation, winnow them down to 12 to 20 variables that are most predictive of the desired
outcome.
In addition, an analytical modeler needs to select historical data that has enough of the “answers”
built in it with a minimal amount of noise. Noise consists of patterns and relationships that have no
business value, such as a person’s birth date and age, which gives a 100% correlation. A data modeler
will eliminate one of those variable to reduce noise. In addition, they will validate their models by
testing them against random subsets of the data which they set aside in advance. If the scores remain
compatible across training, testing and validation data sets, then they know they have a fairly accurate
and relevant model.
Finally, the modeler must choose the right analytical techniques and algorithms or combinations
of techniques to apply to a given hypothesis. This is where modelers’ knowledge of business processes,
project objectives, corporate data and analytical techniques come into play. They may need to try
many combinations of variables and techniques before they generate a model with sufficient predictive
value.
Every analytical technique and algorithm has its strengths and weaknesses, as summarized in the
tables below. The goal is to pick the right modeling technique. So, you have to do as little preparation
and transformation as possible, according to Michael Berry and Gordon Inhofe in their book, “Data
Mining Techniques: For Marketing, Sales and Customer Support.”
Table 1: Analytical Models
In such an environment, it is critical to have a model repository that can track versions, audit
usage and manage a model through its lifecycle. Once an organization has more than one operational
model, it is imperative it implements model management utilities, which most data mining vendors
now support.
Business Analytics is commonly defined as skills, technologies, applications and practices for
continuous iterative investigation of past business performance to gain insight and drive business
planning (Beller and Barnett, 2009). To identify the skillset commonly expected for business analytics
practitioners, we conducted a search of open position announcements using Indeed.com, a specialized
search engine which indexes job postings across numerous company websites as well as job posting
aggregators. We used the keyword “business analytics” to identify open positions in the New York
City metro area. We examined the job listings which were returned by the Indeed search engine and
after iterative evaluation decided to retain a relatively shortlist of positions which: (1) were offered at
large established companies and (2) exemplified the skillset commonly expected in the industry for
similar positions. Our rationale for focusing on the large established corporations is grounded in the
expectation that large companies have more established business processes and more clearly defined
job functions compared to smaller, less established companies (Humphrey, 1988). Our decision to
focus on a limited number of representative positions stems from the observation that while specific
industries and companies may have very distinct jobs requirements, our goal is to identify a common
set of skills that is frequently required across different companies and industries. The positions
selected for our analysis include the following:
Data Visualization Consultant (Accenture)
Data Analytics Manager (Deloitte)
Business Intelligence Analyst (UBS)
Compliance Office Analyst (Citibank)
Data and Analytics Consultant (Accenture)
Loan Operations Business Analyst (Capital One)
14 Business Analytics
Business
Domain
expertise
Figure 5
Our evaluation of the job requirements along the three domains in Figure 6 suggests that applied
statistical skills required by the companies encompass both a theoretical understanding of statistical
methods, as well as practical knowledge of software packages commonly used for statistical analysis –
primarily SAS and R software. The job descriptions commonly require familiarity with regression
modeling techniques. Application of regression analysis requires understanding of inherent
assumptions underlying the regressions, and necessitates foundational statistical knowledge of
distributions, sampling and statistical inference. Though not all job postings explicitly stated this
requirement, we inferred the need for foundational statistical knowledge wherever the position
required regression analysis expertise.
Data mining is a broad concept that encompasses many data model design and analytical
techniques which generally include regression analysis among them (Fayyad, Piatetsky-Shapiro and
Smyth, 1996). In our analysis, we separated regression skills from the more advanced data mining
methodologies, e.g., decision trees, neural networks, support vector machines as well as ensemble
modeling techniques. Further, we also separately evaluated job requirements for text analytical skills,
because analysis of textual data is a unique domain within data mining practice with specialized
expertise related to processing and modeling of textual data. Considering that 80% of the world’s data
today is unstructured, these skills set are becoming extremely important.
The ability to locate, extract and prepare data for analysis is foundational for business analytics in
practice. The required stated technical data management skills among the reviewed job postings span
the range from the basic structured query language (SQL) competency in popular relational database
management systems (RDBMS) to proficiency with large data set analysis leveraging Hadoop
infrastructure. While SQL, RDBMS and data warehousing skills are nearly universally required across
the positions which we reviewed, a growing number of positions also require competency with key-
Introduction to Business Analytics 15
value stores, most commonly exemplified by Hadoop implementations in practice. Data warehousing
job requirements often specifically call for experience with data extraction, transformation, loading
(ETL) and cleaning. Further, two of the eight positions in our sample explicitly required expertise with
Python Programming language as the development platform for performing data processing and
analysis.
Data visualization expertise was nearly universally required by the positions, which we included
in our analysis. Data visualization represents an important area of practice. Virtually, all positions in
our set listed Tableau software as the dominant tool for data visualization, but several positions also
suggested Qlikview as another potential software choice for data visualization. All positions
emphasized the importance of soft skills: effective communication and presentation as well as the
ability to work in groups, highlighting the fact that effective business analytics in practice often
requires group collaboration and effective communication of insights across the enterprise. These
skills become important in influencing the decision to implement the results of analytical exercise/
analytics team.
In addition to specific knowledge of statistical methods and technical skills, every position also
included business domain specific expertise which qualified an ideal job candidate. These
requirements are detailed in Table 3.
Table 3
The volume of data that one has to deal has exploded to unimaginable levels in the past decade,
and at the same time, the price of data storage has systematically reduced. Private companies and
research institutions capture terabytes of data about their users’ interactions, business, social media,
and also sensors from devices such as mobile phones and automobiles. The challenge of this era is to
make sense of this sea of data. This is where big data analytics comes into picture.
Big Data Analytics largely involves collecting data from different sources, mugged it in a way
that it becomes available to be consumed by analysts and finally deliver data products useful to the
organization business.
The process of converting large amounts of unstructured raw data, retrieved from different
sources to a data product useful for organizations, forms the core of Big Data Analytics.
dealing with text, perhaps in different languages normally requiring a significant amount of
time to be completed.
5. Data Mugging: Once the data is retrieved, for example, from the web, it needs to be stored
in an easy-to-use format. To continue with the reviews examples, let’s assume the data is
retrieved from different sites where each has a different display of the data.
Suppose one data source gives reviews in terms of rating in stars. Therefore, it is possible to
read this as a mapping for the response variable y ∈ {1, 2, 3, 4, 5}. Another data source
gives reviews using two arrows system, one for up voting and the other for down voting.
This would imply a response variable of the form y ∈ {positive, negative}.
In order to combine both the data sources, a decision has to be made in order to make these
two response representations equivalent. This can involve converting the first data source
response representation to the second form, considering one star as negative and five stars as
positive. This process often requires a large time allocation to be delivered with good quality.
6. Data Storage: Once the data is processed, it sometimes needs to be stored in a database. Big
data technologies offer plenty of alternatives regarding this point. The most common
alternative is using the Hadoop File System for storage that provides users a limited version
of SQL, known as HIVE Query Language. This allows most analytics task to be done in
similar ways as would be done in traditional BI data warehouses, from the user perspective.
Other storage options to be considered are Mongo DB, Redis and SPARK.
This stage of the cycle is related to the human resources knowledge in terms of their abilities
to implement different architectures. Modified versions of traditional data warehouses are
still being used in large-scale applications. For example, Teradata and IBM offer SQL
databases that can handle terabytes of data; open source solutions such as postgreSQL and
MySQL are still being used for large-scale applications.
Even though there are differences in how the different storages work in the background,
from the client side, most solutions provide a SQL API. Hence, having a good
understanding of SQL is still a key skill to have for big data analytics.
This stage a priori seems to be the most important topic; in practice, this is not true. It is not
even an essential stage. It is possible to implement a big data solution that would be working
with real-time data. So, in this case, we only need to gather data to develop the model and
then implement it in real time. So, there would not be a need to formally store the data at all.
7. Exploratory Data Analysis: Once the data has been cleaned and stored in a way that
insights can be retrieved from it, the data exploration phase is mandatory. The objective of
this stage is to understand the data. This is normally done with statistical techniques and also
plotting the data. This is a good stage to evaluate whether the problem definition makes
sense or is feasible.
8. Data Preparation for Modeling and Assessment: This stage involves reshaping the
cleaned data retrieved previously and using statistical pre-processing for missing values
imputation, outlier detection, normalization, feature extraction and feature selection.
9. Modelling: The prior stage should have produced several data sets for training and testing,
e.g., a predictive model. This stage involves trying different models and looking forward to
solving the business problem at hand. In practice, it is normally desired that the model
would give some insight into the business. Finally, the best model or combination of models
is selected evaluating its performance on a left-out data set.
Introduction to Business Analytics 19
10. Implementation: In this stage, the data product developed is implemented in the data
pipeline of the company. This involves setting up a validation scheme while the data product
is working in order to track its performance. For example, in case of implementing a
predictive model, this stage would involve applying the model to new data and once the
response is available, evaluate the model.
Unstructured
Unstructured data refers to the data that lacks any specific form or structure whatsoever. This
makes it very difficult and time-consuming to process and analyze unstructured data. Email is an
example of unstructured data.
Structured Data Unstructured Data
Characteristics Pre-defined data models No pre-defined data model
Usually text only May be text, images, sound, video or
other formats
Easy to search Difficult to search
Resides in Relational databases Applications
Data warehouses NoSQL databases
Data warehouses
Data lakes
Generated by Humans or Machines Humans or Machines
Typical Airline reservation systems Word processing
applications Inventory control Presentation software
CRM systems Email clients
ERP systems Tools for viewing or editing media
Examples Dates Text Files
Phone numbers Presentation software
Social security numbers Email messages
Credit card numbers Audio files
Customer names Video files
Addresses Images
Product names and numbers Surveillance imagery
Transaction information
20 Business Analytics
Semi-structured
Semi-structured data pertains to the data containing both the formats mentioned above, i.e.,
structured and unstructured data. To be precise, it refers to the data that although has not been
classified under a particular repository (database), yet contains vital information or tags that segregate
individual elements within the data.
The Data
UNSTRUCTURED Landscape
DATA
Flat Text
UNSTRUCTURED
Databases
Being accurate, Big Data combines relevant data from multiple sources to produce highly
actionable insights. Almost 43% of companies lack the necessary tools to filter out irrelevant data,
which eventually costs them millions of dollars to hash out useful data from the bulk. Big Data tools
can help reduce this, saving you both time and money.
Big Data Analytics could help companies generate more sales leads which would naturally mean
a boost in revenue. Businesses are using Big Data Analytics tools to understand how well their
products/services are doing in the market and how the customers are responding to them. Thus, they
can understand better where to invest their time and money.
With Big Data insights, you can always stay a step ahead of your competitors. You can screen the
market to know what kind of promotions and offers your rivals are providing, and then you can come
up with better offers for your customers. Also, Big Data insights allow you to learn customer behavior
to understand the customer trends and provide a highly ‘personalized’ experience to them.
Who is Using Big Data?
The people who are using Big Data know better that, what Big Data is. Let’s look at some such
industries:
Healthcare: Big Data has already started to create a huge difference in the healthcare sector.
With the help of predictive analytics, medical professionals and HCPs are now able to
provide personalized healthcare services to individual patients. Apart from that, fitness
wearables, telemedicine, remote monitoring – all powered by Big Data and AI – are helping
change lives for the better.
Academia: Big Data is also helping enhance education today. Education is no more limited
to the physical bounds of the classroom – there are numerous online educational courses to
learn from. Academic institutions are investing in digital courses powered by Big Data
technologies to aid the all-round development of budding learners.
Banking: The banking sector relies on Big Data for fraud detection. Big Data tools can
efficiently detect fraudulent acts in real-time such as misuse of credit/debit cards, archival of
inspection tracks, faulty alteration in customer stats, etc.
Manufacturing: According to TCS 2013 Global Trend Study, the most significant benefit
of Big Data in manufacturing is improving the supply strategies and product quality. In the
manufacturing sector, Big Data helps create a transparent infrastructure, thereby predicting
uncertainties and incompetencies that can affect the business adversely.
IT: One of the largest users of Big Data, IT companies around the world are using Big Data
to optimize their functioning, enhance employee productivity and minimize risks in business
operations. By combining Big Data technologies with ML and AI, the IT sector is
continually powering innovation to find solutions even for the most complex of problems.
1.6 Questions
I. Essay Type Questions
1. Explain about Business Analytics.
2. Describe categories of Business analytical methods.
3. How Big Data helps in Business Data Analysis?
4. Explain the types of data .
22 Business Analytics