0% found this document useful (0 votes)
234 views20 pages

Business Analytics ASSIGNMENT Questions

The document discusses various aspects of Data Analytics, including its definition, core steps, types, and its relationship with Data Science and Big Data Analytics. It highlights the importance of R as a powerful tool for data analysis, detailing its advantages, visualization capabilities, and popular packages. Additionally, it covers Predictive Data Analytics and Textual Data Analytics, explaining their features and real-world applications.

Uploaded by

Reva Kawatra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
234 views20 pages

Business Analytics ASSIGNMENT Questions

The document discusses various aspects of Data Analytics, including its definition, core steps, types, and its relationship with Data Science and Big Data Analytics. It highlights the importance of R as a powerful tool for data analysis, detailing its advantages, visualization capabilities, and popular packages. Additionally, it covers Predictive Data Analytics and Textual Data Analytics, explaining their features and real-world applications.

Uploaded by

Reva Kawatra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 20

BUSINESS ANALYTICS

ASSIGNMENT:

NAME: REVA

COURSE: B.COM (H)

SECTION: B

ROLL NO. 2022298

SUBJECT: BUSINESS ANALYTICS

SUBMITTED TO: MR. DEVENDER


KUMAR
Ques 1. What do you understand by Data Analytics? How is it associated
with data science and big data analytics?

Ans.1- What is Data Analytics?

Data Analytics is the science of analyzing raw data to extract meaningful


patterns, trends, and insights that help individuals or organizations make informed
decisions. It involves a systematic process of inspecting, cleaning, transforming,
and modeling data. The ultimate goal is to discover useful information, support
decision-making, and often, automate certain insights.

Data Analytics goes beyond just looking at numbers - it’s about turning data into a
story that explains what’s happening in a business, process, or system, and what
actions can be taken to improve outcomes.

Core Steps in Data Analytics:

1. Data Collection – Gathering data from various sources like databases, logs,
sensors, web platforms, etc.
2. Data Cleaning – Removing errors, inconsistencies, and duplicates to
improve data quality.
3. Data Exploration and Transformation – Understanding the structure of
the data, transforming it into usable formats, and performing exploratory
data analysis (EDA).
4. Data Modeling and Analysis – Applying statistical or machine learning
techniques to identify patterns, correlations, or trends.
5. Data Visualization and Interpretation – Presenting results using charts,
graphs, dashboards, or reports to make them understandable and actionable.

Types of Data Analytics:

 Descriptive Analytics: Analyzing historical data to understand what


happened.
 Diagnostic Analytics: Looking into the data to find out why
something happened.
 Predictive Analytics: Using models and forecasts to predict future
outcomes.
 Prescriptive Analytics: Suggesting actions to take for optimal
outcomes.
How is Data Analytics Associated with Data Science?

Data Analytics is a fundamental part of the broader field of Data Science.

Data Science is an interdisciplinary field that combines data analytics,


computer science, mathematics, and domain knowledge to extract insights
from complex and often unstructured data. It not only analyzes the data
but also builds predictive models, automates decisions, and even creates
data-driven products using tools like artificial intelligence (AI) and
machine learning (ML).

Whereas Data Analytics often answers specific questions using structured


data (e.g., sales trends, customer behavior), Data Science goes deeper to:

 Build and train machine learning algorithms,


 Create predictive models,
 Handle unstructured data like images, audio, and text,
 Develop algorithms for automation and recommendation systems.

In essence, data analytics is one of the tools or skill sets that data scientists
use regularly.

How is Data Analytics Related to Big Data Analytics?

Big Data Analytics is a specialized branch of data analytics that focuses on


analyzing large, complex, and fast-moving datasets—commonly referred to
as "Big Data."

Characteristics of Big Data:

 Volume: Massive amounts of data generated from various sources


like social media, IoT devices, sensors, and transaction records.
 Velocity: The speed at which data is generated and needs to be
processed.
 Variety: The many different formats of data—structured, semi-
structured, and unstructured.
 Veracity: The trustworthiness and quality of the data.
 Value: The potential value that can be extracted from the data.

Big Data Analytics requires advanced tools and frameworks such as:

 Apache Hadoop and Apache Spark for distributed data processing,


 NoSQL databases like MongoDB or Cassandra,
 Cloud computing platforms like AWS, Azure, and Google Cloud,
 Real-time data streaming tools like Apache Kafka.

Where traditional data analytics tools might struggle with performance or


scalability, big data analytics tools are designed to manage data at scale.

Conclusion:

 Data Analytics is the foundation—it helps make sense of data and extract insights to support
decisions.
 It is a core function within Data Science, which is a broader, more holistic field focused on not
just analysis, but also prediction, automation, and innovation.
 Big Data Analytics applies data analytics techniques to extremely large and complex datasets
that cannot be handled with traditional methods.
 All three are closely interconnected and often work hand-in-hand in modern data-driven
environments. Whether it’s a business analyzing customer behavior, a healthcare provider
predicting disease outbreaks, or a social media platform recommending content—data
analytics is at the heart of making data useful.

Ques 2. “R is used extensively for data analysis”. In the light of this statement,
discuss some of the advantages of using R for data analysis.
Ans 2. The statement “R is used extensively for data analysis” is very accurate
—R has long been one of the most popular tools in the field of data
analysis and statistics. Originally developed by statisticians for
statisticians, R has evolved into a powerful programming language and
environment specifically designed for data manipulation, statistical modeling,
and visualization.

Let’s explore some of the key advantages of using R for data analysis :

1. Designed for Statistical Computing:

R was built from the ground up for statistics and data analysis. It provides a vast
number of functions for statistical techniques like:

 Linear and nonlinear modeling,


 Time-series analysis,
 Classification and clustering,
 Hypothesis testing and regression,
 ANOVA, chi-square tests, and more.

This makes it ideal for anyone working with data in a research, academic,
or business setting.

2. Rich Ecosystem of Packages:

R has a huge number of packages available through:

 CRAN (Comprehensive R Archive Network),


 Bioconductor (for bioinformatics),
 GitHub (community-developed tools).

These packages extend R's functionality to handle a wide variety of data


analysis tasks—from data cleaning and mining to machine learning and
reporting. Popular ones include:

 dplyr and tidyr for data manipulation,


 ggplot2 for data visualization,
 caret and randomForest for machine learning,
 shiny for creating interactive web apps.
3. Excellent Data Visualization Capabilities:

R excels at data visualization, especially with the ggplot2 package, which is


based on the "Grammar of Graphics". It allows users to build complex,
customizable, and publication-ready charts with relatively simple code.

From bar charts and histograms to heatmaps, boxplots, and even 3D


visualizations, R makes it easy to create visuals that bring data to life.

4. Strong Community Support:

R has a large, active, and welcoming community. This means:

 Plenty of online forums (like Stack Overflow and RStudio


Community),
 Extensive documentation and tutorials,
 Regularly updated packages and contributions from researchers and
practitioners worldwide.

Whether you're a beginner or an expert, chances are someone has already


solved the problem you’re facing—and shared their solution.

5. Open Source and Free:

R is completely free to download, use, and modify. This is especially


beneficial for:

 Students and academic researchers with limited budgets,


 Startups and businesses that want to avoid costly licenses.

As an open-source language, it’s constantly evolving with contributions


from users around the world.

6. Integration and Compatibility:

R can easily integrate with:

 Databases like MySQL, PostgreSQL, and SQLite,


 Big Data tools like Hadoop and Spark,
 Other languages like Python, C++, and Java,
 Excel and CSV files, making it highly flexible in handling different
types of data sources.

You can even embed R in web applications (using shiny) or use it in reports
(via RMarkdown) that combine code, text, and visuals.

7. Ideal for Reproducible Research:

With tools like R Markdown, knitr, and bookdown, R makes it easy to create
documents that include code, analysis, and narrative text—all in one
place. This is essential for ensuring reproducibility, which is a core
principle in both academic and industry data science projects.

8. Versatility in Applications:

While R is best known for statistical analysis, it is also used in:

 Machine learning,
 Data mining,
 Text analysis and natural language processing,
 Bioinformatics,
 Finance and economics,
 Marketing analytics, and more.

This versatility adds to its value as a general-purpose data analysis tool.

Conclusion:

R remains a top choice for data analysis because of its specialized features, extensive libraries, and
visualization capabilities. Its design makes complex statistical tasks much easier, and its ecosystem
supports everything from small-scale data wrangling to large-scale predictive modeling.

If your focus is statistical modeling, visualization, or reproducible research, R is often the go-to
language for professionals and academics alike.

Ques 3. What is data visualization. Describe the various types of tools and
packages that can be used to create powerful visualization of data in R.
Ans 3. What is Data Visualization?

Data Visualization is the graphical representation of information and data.


By using visual elements like charts, graphs, and maps, data visualization
tools make complex data more accessible, understandable, and usable.

It helps in:

 Identifying trends and patterns,


 Communicating insights clearly,
 Supporting decision-making with visuals,
 Making large datasets easier to interpret.

In short, data visualization transforms raw data into visuals that tell a
story.

Data Visualization in R:

R is one of the most powerful languages for data visualization. It provides


a wide range of libraries and packages to create simple to highly
customized, interactive, and publication-quality graphics.

Here are some of the most popular and powerful tools and packages in R
for data visualization:

1. Ggplot2:

 Probably the most popular and powerful data visualization package


in R.
 Based on The Grammar of Graphics—a consistent and layered
approach to creating visuals.
 Used for making static, highly customizable, and publication-quality
graphics.

Examples of charts you can make with ggplot2:

 Bar charts
 Line graphs
 Histograms
 Scatter plots
 Boxplots
 Density plots
 Faceted/multi-panel plots

In R:-
CopyEdit
library(ggplot2)
ggplot(data = mtcars, aes(x = mpg, y = hp)) +
geom_point() +
theme_minimal()

2. Lattice:

 Great for creating trellis graphics (multi-variable and multi-panel


plots).
 Suitable for more statistical or scientific visualizations.
 Built-in functions simplify multi-dimensional data visualization.

In R:-

CopyEdit
library(lattice)
xyplot(mpg ~ wt | cyl, data = mtcars)

3. Plotly:

 Used for interactive web-based visualizations.


 Works with both base R plots and ggplot2.
 Allows users to zoom, hover, and click on plot elements.

In R:-

CopyEdit
library(plotly)
p <- ggplot(mtcars, aes(x = wt, y = mpg)) + geom_point()
ggplotly(p)

4. Base R plotting system:

 R comes with built-in basic plotting capabilities.


 Suitable for quick and simple plots without extra libraries.
 Not as visually polished as ggplot2, but still powerful for small tasks.

In R:-

CopyEdit
plot(mtcars$mpg, mtcars$hp, main="MPG vs HP")
5. Highcharter:

 A wrapper for the popular Highcharts JavaScript library.


 Used for creating interactive charts that look great on the web.
 Excellent for dashboards and web apps.

In R:-

CopyEdit
library(highcharter)
highchart() %>%
hc_add_series(data = c(3, 5, 7, 9), type = "line")

6. Leaflet:

 Specialized for interactive maps and geospatial data visualization.


 Allows you to easily create maps with layers, markers, and pop-ups.

In R:-

CopyEdit
library(leaflet)
leaflet() %>%
addTiles() %>%
addMarkers(lng = -0.09, lat = 51.505, popup = "London")

7. Esquisse:

 A drag-and-drop GUI tool built on top of ggplot2.


 Lets you create complex visualizations without writing code.
 Great for beginners or exploratory analysis.

In R:-

CopyEdit
# Install and load
# esquisse::esquisser() launches the GUI

8. RShiny + RMarkdown:

While not packages specifically for plotting, they allow you to present your
visualizations interactively:

 shiny: Turn your R visualizations into interactive web applications.


 rmarkdown: Combine code, narrative text, and visualizations in one
document.
Ques 4. What is predictive Data Analytics. Discuss its features using suitable
examples.

Ans 4. What is Predictive Data analytics?

Predictive Data Analytics is a branch of data analytics that focuses on


forecasting future outcomes based on historical data, patterns, and
statistical modeling. It uses techniques such as machine learning, data
mining, and predictive modeling to anticipate what might happen in the
future.

At its core, predictive analytics answers the question:


👉 “What is likely to happen next?”

It doesn't just explain the past or present—it foresees potential trends,


behaviors, or events, helping organizations make proactive, data-driven
decisions.

Key Features of Predictive Data Analytics:

1. Data-Driven Forecasting,

Predictive analytics uses historical data (e.g., past purchases, weather,


website clicks) to detect patterns that can predict future behavior.

Example:
An e-commerce platform might use predictive analytics to forecast which
products a customer is likely to buy based on their previous browsing and
purchase history.

2. Machine Learning and Statistical Models,

It relies heavily on advanced algorithms like:

 Regression analysis (linear, logistic),


 Decision trees,
 Neural networks,
 Random forests,
 Time-series forecasting.

These models "learn" from the data and improve over time.
Example:
A bank might use a logistic regression model to predict the likelihood of a
customer defaulting on a loan.

3. Real-Time or Near Real-Time Processing,

Many predictive analytics systems can process incoming data in near real-
time, enabling businesses to respond instantly.

Example:
A fraud detection system in a bank analyzes transactions in real-time to
predict if a purchase might be fraudulent.

4. Risk Assessment and Scoring,

Predictive analytics is widely used for risk modeling, where it assigns scores
or probabilities to assess risk levels.

Example:
In insurance, predictive analytics can assess which applicants are high-risk
based on demographic and behavioral data.

5. Behavior Prediction,

It helps in understanding user behavior, preferences, or potential actions—


essential for personalization and marketing.

Example:
Netflix uses predictive models to recommend shows and movies you’re
likely to watch based on your viewing history and ratings.

6. Scenario Modeling / "What-If" Analysis,

Predictive models allow businesses to simulate different scenarios and


assess potential outcomes.
Example:
A supply chain manager might simulate how changing shipping providers
could impact delivery times and costs.

7. Automated Decision-Making,

In many systems, predictive analytics feeds into automated processes, such


as:

 Dynamic pricing,
 Personalized marketing,
 Inventory management.

Example:
Amazon adjusts product prices based on predicted demand and
competitor activity.

Real-World Applications of Predictive Analytics:-

Industry Application
Retail: Predict customer churn, optimize pricing & inventory
Healthcare: Forecast disease outbreaks, hospital readmission
risks
Finance: Credit scoring, fraud detection
Marketing: Targeted ad campaigns, customer segmentation
Manufacturing: Predictive maintenance, supply chain optimization
Transportation: Route optimization, demand forecasting

Conclusion:

Predictive Data Analytics is not just about knowing what happened, but about knowing what’s coming.
With access to quality data and the right models, it empowers organizations to plan ahead, reduce
risks, improve customer satisfaction, and stay competitive.
Ques 5. What is Textual Data Analytics? Describe the major application of
textual data analysis.

Ans 5. What is Textual Data Analytics?

Textual Data Analysis, also known as Text Analytics or Text Mining, is the process
of extracting useful insights, patterns, and meaningful information from
unstructured text data. This type of analysis helps computers understand human
language and turn messy text into structured, analyzable data.
Since a large portion of data generated today is textual (emails, social media posts,
customer reviews, chat logs, etc.), text analysis has become a critical area in data
science and artificial intelligence.

Key Components of Textual Data Analysis:

1. Text Preprocessing:
o Removing stop words (like "the", "is", "and")
o Tokenization (breaking text into words or phrases)
o Stemming/Lemmatization (reducing words to root form)
o Removing punctuation, numbers, and irrelevant symbols
2. Text Representation:
o Bag-of-Words (BoW): Represents text as a set of word frequencies.
o TF-IDF (Term Frequency-Inverse Document Frequency): Weights
words based on importance.
o Word Embeddings (Word2Vec, GloVe): Converts words into
vectors capturing semantic meaning.
3. Natural Language Processing (NLP):
o Part-of-Speech tagging
o Named Entity Recognition (NER)
o Syntax and grammar parsing
4. Text Classification & Clustering:
o Grouping similar documents or classifying texts into categories
(e.g., spam vs. not spam)
5. Sentiment Analysis:
o Detecting the emotional tone behind the text (positive,
negative, neutral)

Major Applications of Textual Data Analysis:

1. Sentiment Analysis,

Analyzing customer opinions or public sentiment from reviews, tweets, or


feedback.

Example:
A company analyzes social media posts to gauge how people feel about a
new product launch.
2. Spam Detection,

Identifying and filtering unwanted or harmful messages like spam emails


or phishing attempts.

Example:
Email services use text analysis to detect keywords and patterns in spam
emails.

3. Customer Feedback Analysis,

Understanding what customers are saying about products or services


through surveys, reviews, or help desk chats.

Example:
An e-commerce platform uses text analysis to find common complaints in
product reviews.

4. Chatbots and Virtual Assistants,

Used to understand and respond to human queries in natural language.

Example:
AI chatbots in banking apps analyze user messages to offer help with
account-related questions.

5. Topic Modeling,

Uncovering hidden themes or topics in large volumes of text data.

Example:
A news agency uses topic modeling to categorize thousands of news articles
automatically.
6. Healthcare and Medical Records Analysis,

Extracting information from clinical notes, prescriptions, and medical


records.

Example:
Hospitals use text analysis to identify trends in symptoms and diagnoses
from patient records.

7. Legal and Policy Document Analysis,

Analyzing contracts, case laws, or policy documents to detect key clauses


or compliance issues.

Example:
A law firm uses text mining to search thousands of documents for relevant
case precedents.

8. Fraud Detection,

Spotting unusual patterns or language in insurance claims, loan


applications, etc.

Example:
Insurance companies use text analysis to flag suspicious language in claim
descriptions.

Conclusion:

Textual Data Analysis transforms unstructured text into actionable insights. In an age where text is
one of the most common forms of data, its applications span across industries—from marketing and
healthcare to finance and law. By leveraging text analysis, organizations can better understand human
communication, automate decision-making, and improve customer experience.

You might also like