0% found this document useful (0 votes)
14 views11 pages

DBB1202

The document provides an overview of statistics, including its functions such as data collection, organization, and analysis, as well as its applications in various fields like business and healthcare. It also discusses different types of data classification, including qualitative, quantitative, structured, unstructured, semi-structured, time series, and spatial data. Additionally, it covers methods for calculating correlation coefficients and the components of time series analysis, including trends, seasonality, and cyclic patterns.

Uploaded by

Tamannah
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views11 pages

DBB1202

The document provides an overview of statistics, including its functions such as data collection, organization, and analysis, as well as its applications in various fields like business and healthcare. It also discusses different types of data classification, including qualitative, quantitative, structured, unstructured, semi-structured, time series, and spatial data. Additionally, it covers methods for calculating correlation coefficients and the components of time series analysis, including trends, seasonality, and cyclic patterns.

Uploaded by

Tamannah
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 11

NAME TAMANNA RASTOGI

ROLL NO. 2314511771


PROGRAM BACHELOR OF BUSINESS
ADMINISTRATION (BBA)
SEMESTER III
COURSE NAME QUANTITATIVE TECHNIQUES FOR
MANAGEMENT
COURSE CODE DBB2102
SET-1

ANSWER 1 -: Statistics is a branch of mathematics that deals with collecting, analyzing,


interpreting, presenting, and organizing data. It plays a crucial role in various fields, including
science, business, economics, social sciences, and healthcare helping to make informed
decisions, identify patterns, and draw conclusions.
One of the primary functions of statistics is data collection. Before any analysis can occur,
relevant and reliable data must be gathered. This can be done through various methods such as
surveys, experiments, and observational studies. The data collection process is crucial because
the quality and accuracy of the data directly affect the validity of the analysis.
Once the data is collected, the next function of statistics is data organizing. This involves sorting
and summarizing the data to make it more understandable. Techniques like frequency
distributions, tables and charts (such as histograms and pie charts) are used to prevent data in a
way that highlights key trends and patterns. This step makes large datasets easier to interpret and
paves the way for deeper analysis.
After organizing the data, statistical analysis begins. Descriptive statistics such as mean, median,
mode, and standard deviation, are used to summarize and describe the characteristics of data.
These measures help in understanding the central tendency and variability within a dataset. For
example, the mean gives an average value, while the standard deviation shows how spread out
the data is.
Another important aspect of statistical analysis is inferential statistics. This branch involves
making predictions or generalizations about a population based on a sample. It uses probability
theory to estimate population parameters, test hypotheses, and make inferences. Inferential
statistics helps determine the likelihood of certain outcomes, assess the reliability of predictions,
and test the validity of assumptions.
In additions to these core functions, statistics also helps with decision-making by providing tools
for uncertainty and risk assessment. In business, for example, statistical methods are used for
market research, quality control and forecasting future trends. In healthcare, statistics play a key
role in analyzing clinical traits, determining the effectiveness of treatments and studying disease
prevalence.
Lastly, statistics aids in hypothesis testing and drawing conclusions about relationships between
variables. For example, it can help assess whether there is a significant difference between two
groups or if one variable influence another.
Overall, the function of statistics is to provide a structured framework for analyzing data,
drawing meaningful conclusions, and supporting decision making in the face of uncertainty.
ANSWER 2 -: Classification of data refers to the process of categorizing data into different
groups or classes based on certain characteristics or features. This classification helps in
organizing data for better analysis, interpretation, and decision-making. Data can be classified
into various types, depending on its nature and the type of analysis being performed.
1. QUALITATIVE DATA (categorial data)
It refers to non-numeric information that describes qualities or characteristics. It can be
further divided into two types -:
 Nominal data – this type of data represents categories with no inherent order or
ranking. Example include colors (red, blue, green), gender (male, female), or
types of animals (cat, dog, bird).

 Ordinal data – it involves categories that have a meaningful order or ranking, but
the intervals between categories are not necessarily consistent. Examples include
education level (high school, college, graduate) or customer satisfaction ratings
(poor, average, excellent).

2. QUANTITATIVE DATA
It consists of numerical values and can be measured and quantified. It is typically used
for mathematical analysis and statistical modeling. This type can be split into -:
 Discrete data – this data is countable and can consists of distinct separate values.
For example, the number of students in a class, the numbers of cars in a parking
lot, or the number of defects in a product.

 Continuous data – it can take any value within a given range and is measurable.
For instance, height, weight, temperature, or time. These values are often
measured with precision, and there can be infinite possibilities within a range.

3. STRUCTURED DATA
It is highly organized and easily searchable because it is stored in a predefined format,
such as databases or spreadsheets. Each data element is stored in a fixed field within a
record. Example include customer information (name, address, phone number) or
financial data (sales, figures, profit margins).

4. UNSTRUCTURED DATA
It is not organized in a predefined manner, making it harder to search or analyze. It
typically includes text-heavy formats or multimedia. Examples include, emails, social
media posts, videos, images, and audio files.

5. SEMI-STRUCTURED DATA
It lies between structured and non-structured data. It does not fit neatly into tables, but it
still contains some level of organization through tags or markers. Examples include XML
files, JSON files, or NoSQL databases.
6. TIME SERIES DATA
This type of data is collected at regular intervals over time and is useful for tracking
trends or changes over a period. Examples include stock prices, weather data, or traffic
data.

7. SPATIAL DATA
It represents information about physical location and their attributes. It includes
geographic data, such as coordinates, maps and satellite images.

ANSWER 3 (a) -: To calculate the mean of a frequency distribution, we use the formula -:
𝛴(𝑓.𝑋)
Mean =
𝛴𝑓
Where:

 𝑓 is the frequency
 X is the value of the mark
 ∑ (𝑓 . X) is the sum of the product of frequency and marks.
 ∑ 𝑓 is the total frequency.
Sum of the products of frequency and marks
Marks X Frequency 𝑓 Product
10 8 10*8 =80
20 12 20*12=240
30 20 30*20= 600
40 10 40*10=400
50 7 50*7= 350
60 3 60*3= 180

∑ (𝑓. 𝑋) = 80 + 240 + 600 + 400 + 350 + 180 = 1850


∑ 𝑓 = 8 + 12 + 20 + 10 + 7 + 3 = 60

1850
Mean = = 30.83
60
So, the mean of given frequency distribution is 30.83
b) Calculating the cumulative frequency from the given data -:
Size (X) Frequency (f) Cumulative frequency (CF)
4 10 10
4.5 18 28
5 22 50
5.5 25 75
6 40 115
6.5 15 130
7 10 140
7.5 8 148
8 7 155

Total number of observations (N) = Sum of all the frequencies


𝑁 = 10 + 18 + 22 + 25 + 40 + 15 + 10 + 8 + 7 = 155
To find the first quartile (Q1) -:
The first Quartile (Q1) corresponds to the 25th percentile, which is the value that divides the
lower 25% of the data. To find Q1 using the following formula -:
𝑁 155
𝑄1 = = = 38.75
4 4

This means that Q1 corresponds to the position 38.75 in the cumulative frequency distribution.
To find this, we need to look for the size where the cumulative frequency first exceeds 38.75
From the cumulative frequency table
 The cumulative frequency for size 4 is 10.
 The cumulative frequency for size 4.5 is 28.
 The cumulative frequency for size 5 is 50.
Since 38.75 lies between 28 and 50, Q1 is between 4.5 and 5.
Using linear interpolation to find the exact value of Q1.
𝑁
− 𝐶𝐹
𝑄1 = 𝐿 + ( 4 )×ℎ
𝑓

Where:

 𝐿 is the lower boundary of the group containing Q1 (5 in this case).


 𝐶𝐹 is the cumulative frequency before the group (28 for the size 4.5).
 𝑓 is the frequency of the group (22 for size 5).
 ℎ is the class width (0.5, since the classes are 4 to 4.5, 4.5 to 5 etc.)

38.75 − 28
𝑄1 = 5 + ( ) × 0.5
22
10.75
𝑄1 = 5 + ( ) × 0.5
22
𝑄1 = 5 + (0.4886) × 0.5
𝑄1 = 5 + 0.2443 = 5.2443
Q1 ≈ 5.2443

To find the third quartile (Q3) -:


The third Quartile (Q3) corresponds to the 75th percentile, which is the value that divides the
lower 75% of the data. To find Q3 using the following formula -:
3𝑁 3 ×155
𝑄3 = = = 116.25
4 4

This means that Q3 corresponds to the position 116.25 in the cumulative frequency distribution.
To find this, we need to look for the size where the cumulative frequency first exceeds 116.25
From the cumulative frequency table

 The cumulative frequency for size 6 is 115.


 The cumulative frequency for size 6.5 is 130.
Since 116.25 lies between 115 and 130, Q3 is between 6 and 6.5.
Using linear interpolation to find the exact value of Q3.
3𝑁
− 𝐶𝐹
𝑄3 = 𝐿 + ( 4 )×ℎ
𝑓

Where:

 𝐿 is the lower boundary of the group containing Q3 (6 in this case).


 𝐶𝐹 is the cumulative frequency before the group (115 for the size 6).
 𝑓 is the frequency of the group (40 for size 6).
 ℎ is the class width (0.5)

116.25 − 115
𝑄3 = 6 + ( ) × 0.5
40
1.25
𝑄3 = 6 + ( ) × 0.5
40
𝑄3 = 5 + (0.03125) × 0.5
𝑄3 = 5 + 0.015625 = 6.0156
Q3 ≈ 6.0156
So, the Quartile one (Q1) is 5.2443 and Quartile three (Q3) is 6.0156.
SET-2

ANSWER 1 -: The coefficient of correlation is a statistical measure that quantifies the strength
and direction of the relationship between two variables. It ranges from -1 to + 1, where +1
indicates a perfect positive correlation, -1 indicates a perfect negative correlation, and 0 indicates
no relationship. The most common type in Pearson’s correlation coefficient, which measures
linear relationship in continuous data. Other methods like Spearman’s rank and Kendall’s tau are
used for ordinal or non-linear relationships. The coefficient helps in understanding how changes
in one variable might predict changes in another.
There are several methods to calculate the correlation coefficient. The most common ones are –
1. Pearson’s Correlation Coefficient
Pearson’s Correlation Coefficient (r) measures the linear relationship between two
continuous variables. It is the most commonly used method and assumes that the
relationship between the variables is linear and that the data follows a normal
distribution.
The formula of Pearson’s correlation coefficient is –
𝑛 (𝛴 𝑋𝑌) − (𝛴𝑋)(𝛴𝑌)
𝑟=
√[𝑛 𝛴 𝑋 2 − (𝛴 𝑋)2 ][𝑛 𝛴 𝑌 2 − (𝛴 𝑌)2 ]

Where:
 𝑛 is the number of data points.
 𝛴 𝑋𝑌 is the sum of the product paired scores.
 𝛴𝑋 and 𝛴Y are the sums of the values of the variables X and Y respectively.
 𝛴 𝑋 2 and 𝛴𝑌 2 are sums of the squares of X and Y, respectively.
This formula calculates the degree to which the values of X and Y are related linearly. The closer
the result is to +1 or -1, the stronger the linear relationship.

2. Spearman’s Rank Correlation Coefficient


Spearman’s Rank Correlation Coefficient is used when the data is not necessarily linear,
or when the variables are ordinal (ranked). It measures the strength and the direction of
the monotonic relationship between two variables. A monotonic relationship means that
as one variable increases, the other either always increases or always decreases, but not
necessarily at a constant rate.
The formula for Spearman’s Rank correlation coefficient is
6 𝛴 𝑑2
𝑝 = 1−
𝑛(𝑛2 − 1)
Where:

 𝑑 is the difference between the ranks of the period values.


 𝑛 is the number of data points.
The value of 𝑝 ranges from -1 to +1, where a positive value indicates a positive monotonic
relationship, a negative value indicates a negative monotonic relationship, and 0 indicates no
monotonic relationship.
3. Kendall’s Tau Coefficient
Kendall’s tau is another rank-based correlation measure, used for ordinal data. It
compares the number of concordant and discordant pairs of values to calculate the
correlation. While both Spearman’s rank and Kendall’s tau assess the monotonic
relationship, Kendall’s tau is generally considered more robust in small datasets.

ANSWER 2 -: Time series analysis is a statistical technique used to analyze data points collected
or recorded at specific time intervals. This type of analysis is useful for forecasting, identifying
trends, and understanding patterns in data that change over time. Time series data can be
collected on anything that evolves over time, such as stock prices, weather data, or economic
indicators. The key components of time series analysis are -:
1. Trend – The trend component represents the long-term movement or direction in the data
over a significant period. It indicates whether the data points are increasing, decreasing or
remaining constant. For example, the gradual increase in global temperatures over the
past few decades would be considered a trend. Trends can be linear (a steady upward or
downward movement) or nonlinear (such as exponential growth).

2. Seasonality – This component reflects periodic fluctuations in the data that occur at
regular intervals, often related to specific time periods like months, quarters, or seasons.
These patterns are typically driven by factors like climate, holidays, and business cycles.
For example, retail sales might experience higher demand during the holiday season each
year. Identifying seasonal patterns helps in forecasting and planning.

3. Cyclic patterns – cyclical variations occur over irregular intervals and are generally
linked to economic or business cycle. Unlike seasonality, cycles do not have a fixed
period and can be influenced by broader factors like economic booms or recissions. For
example, the growth and decline of a country’s economy could lead to cyclical changes
in employment rates or consumer spending.

4. Noise (irregular component) – it is also known as the residual or error component, refers
to random variations or irregularities in the data that cannot be explained by trend,
seasonality, or cycles. These are often short-term fluctuations or anomalies that arise from
unpredictable factors such as accidents, natural disasters, or sudden market shocks. Noise
typically represents the “background” variability in the data.

5. Level – The level is the baseline value around which the time series fluctuates. It
indicates the overall magnitude of the data series after removing trends, seasonality and
cycles. It’s important in understanding the average value of the series and in detecting
whether the time series is generally high, low, centered at some value.

6. Stationarity – A stationary time series has statistical properties (like mean, variance, and
autocorrection) that do not change over time. Many time series analysis techniques, such
as ARIMA models, assume that the data is stationary. If a series is not stationary, it may
need to be transformed, often by differencing or detrending, before analysis.

ANSWER 3 (a) -: To construct an index number for 2015 with 2014 as the base year, we will use
the Laspeyres Index formula. The formula is -:
𝛴(𝑃2015 × 𝑄2014 )
𝐼= ( ) × 100
𝛴(𝑃2014 × 𝑄2014 )

Where:
 𝑃2014 is the price in 2014,
 𝑃2015 is the price in 2015,
 𝑄2014 is the quantity (or weight) of the commodity in 2014
Since no quantity is provided, we can assume equal quantities for each commodity.
Price relative for each commodity is calculated as
𝑃
Price relative = 𝑃2015 × 100
2014

For each commodity


95
 Commodity A = 90 × 100 = 105.56

60
 Commodity B = 40 × 100 = 150

110
 Commodity C = × 100 = 122.22
90

35
 Commodity D = 30 × 100 = 116.67

Now, the average of these price relatives to get the overall index –
105.56+150+122.22+116.67
Index = = 123.11
4

The index number for 2015, taking 2014 as the base year, is 123.11.

(b) i) PARAMETER -: A parameter is a variable used to define a function, method or procedure


in programming and mathematics. It acts as an input to a function, allowing it to process values
or changes its behavior depending on the given arguments. In programming, parameters are
specified in the function, definition, and the actual values passed during function calls are called
arguments. Parameters can be of various types, such as integers, strings, or objects, depending on
the function’s requirements. They make functions reusable and flexible by engaging them to
operate on different inputs. In mathematics, parameters often represent constants in equations
that can be adjusted to change the outcome of a model or equation. Overall, parameters are key
to customizing functionality and enhancing the adaptability of code or mathematical models.

ii) ESTIMATOR -: An estimator is a statistical tool used to estimate the value of an unknown
population parameter based on sample data. It is a function or formula that provides an
approximation of parameters such as the population mean, variance, or proportion. For instance,
the sample mean can be used as an estimator for the population mean. Estimators are categorized
as either point estimators, which provide a single value estimate, or interval estimators, which
give a range of possible values. A good estimator should be unbiased, meaning its expected value
equals the true parameter, and consistent, meaning it produces more accurate estimates as the
sample size increases. Estimators are central to statistical inference, aiding in decision-making
and predictions based on data. Their reliability and accuracy depend on the sample size and
method used.

You might also like