0% found this document useful (0 votes)
23 views24 pages

Unit 3 KMS

Uploaded by

khushirpatel17
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views24 pages

Unit 3 KMS

Uploaded by

khushirpatel17
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 24

SUBJECT

Statistical Method &


Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Unit 3 Applied Statistics


1.1 Moments, Expectation, Dispersion, Skewness, Kurtosis, Expected Value Of Two
Dimensional Random Variable
MOMENTS:
Moments in statistics are quantitative measure (or a set of statistical parameters) that describes
the specific characteristics of a probability distribution. In simple terms, the moment is a way
to measure how spread out or concentrated the number in a dataset is around the central value,
such as the mean.
Note:
 Mathematically, for any random variable X, the moments of X are defined as the expected
values of X.
Example: E[X], E[X2], E[X3],…
 Moments in statistics are closely related to the concepts of the moment in Physics.
 If a function represents the physical density of the body, then the
 Zero moment is the total mass.
 First moment divided by the total mass is the centre of mass.
 The second moment is rotational inertia.
 In statistics, the first four moments are Mean, Variance, Skewness, and Kurtosis.
 Each of these four moments gives information about the dataset, such as central location,
dispersion, asymmetry, and outliers
Now, let’s discuss the first four moments in detail. Each of these four moments
The rth moment of a frequency distribution about mean X̅ denoted by μr is defined by
̅ )𝒓
𝜮𝒇(𝑿 −𝑿
𝝁𝒓 = , 𝐖𝐡𝐞𝐫𝐞 𝐫 = 𝟎, 𝟏, 𝟐, 𝟑, … … (𝟏)
𝒏
Thus μ1 stands for first moment about mean, μ2 stands for second moment about mean and so
on are given by
𝜮𝒇(𝑿 − 𝑿 ̅) 𝜮𝒇(𝑿 − 𝑿 ̅ )𝟐 𝜮𝒇(𝑿 − 𝑿̅ )𝟑 ̅ )𝟒
𝜮𝒇(𝑿 − 𝑿
𝝁𝟏 = , 𝝁𝟐 = , 𝝁𝟑 = , 𝝁𝟒 =
𝒏 𝒏 𝒏 𝒏
Moments about any arbitrary origin:
The rth moment about any arbitrary point (value) a denoted by μ′r is defined by
𝜮𝒇(𝒙 − 𝒂)𝒓
𝝁′𝒓 = ,𝐖𝐡𝐞𝐫𝐞 𝐫 = 𝟏, 𝟐, 𝟑, … … . (𝟐)
𝒏
 Particular cases:
 Putting r = 0 in (2), we get
𝛴𝑓
𝜇0′ = = 1 i. e. 𝜇0′ = 1
𝑛
 Putting r = 1 in (2), we get
𝜮𝒇(𝒙 − 𝒂) 𝜮𝒇𝒙 𝜮𝒇
𝝁′𝟏 = = − ̅ − 𝒂 OR 𝑿 = 𝝁′𝟏 + 𝒂 … (3)
. 𝒂 = 𝑿 − 𝒂, i.e. 𝝁′𝟏 = 𝑿
𝒏 𝒏 𝒏

Relation between Moment (about mean and about origin)


 𝝁𝟏 = 𝟎 ( 𝒂𝒍𝒘𝒂𝒚𝒔)

DEPARTMENT OF COMPUTER APPLICATION Page | 1


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

 𝝁𝟐 = 𝝁′𝟐 − 𝝁′𝟏 𝟐
 𝝁𝟑 = 𝝁′𝟑 − 𝟑𝝁′𝟏 𝝁′𝟐 + 𝟐𝝁′𝟏 𝟑
 𝝁𝟒 = 𝝁′𝟒 + 𝟔𝝁′𝟏 𝟐 𝝁′𝟐 − 𝟒𝝁′𝟏 𝝁′𝟑 − 𝟑 𝝁′𝟏 𝟒

Four Keys of Moments in Statistics


First Moment – Mean
The first moment is referred to as mean. Mean is a measure of central tendency that represents
the average value of the dataset. It is defined as the ratio of the sum of all observations to the
total number of observations.
i.e.,
∑𝒙
Mean (µ1) = 𝒏 𝒊
Where,
∑ 𝒙𝒊 : Sum of all data points (xᵢ)
n: Number of data points

Example: Let the height of 5 students in a class is 150cm, 155cm, 160m, 165cm, and
180cm. Find the mean height of the class.
Solution:-
Sum of height of all students = 150 + 155 + 160 + 165 + 180 = 810
Total number of students = 5
810
= Mean = = 162.
5
Hence, the mean height of the class is 162 cm.

Second Moment – Variance


Variance (µ2)
Variance is the measure of dispersion, referred to as the second moment in statistics. It is the
average squared difference of data points from the mean. The higher the variance value, the
greater the variability (i.e., data points are far away from the mean) within the dataset, and vice
versa.
Variance measures how far each data point in the dataset is from the mean.
Formula of Variance
∑(𝐱ᵢ – µ)𝟐
Variance (σ²) = 𝒏
Where,
∑(𝐱ᵢ – µ)𝟐 : Sum of the squared difference between each data point (xᵢ) and the mean (µ)
n: Number of data points
Standard Deviation
It is defined as the square root of variance. It compares the spread of two different datasets with
approximately the same mean.
Formula of Standard Deviation

DEPARTMENT OF COMPUTER APPLICATION Page | 2


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah
𝟏
Standard Deviation = (𝒗𝒂𝒓𝒊𝒂𝒏𝒄𝒆)𝟐

Third Moment – Skewness


Skewness is the third moment, which is a measure of lack of symmetry, i.e., measures the
deviation of the given distribution of a random variable from a symmetric distribution. In
simple terms, skewness means the lack of straightness or symmetry.
Formula of Skewness

∑(𝐱ᵢ – µ)𝟑
µ3 = Σ = 𝒏
∑(𝐱ᵢ – µ)𝟑 µ𝟑 𝟐
Skewness (γ) = 𝟑
=
𝒏∗µ𝟐 µ𝟐 𝟑

Where, ∑(𝐱ᵢ – µ)𝟑 : Sum of the cubed difference between each data point (xᵢ) and the mean
(µ).
n: Number of data points.

There are two types of skewness:


Positive Skewness: In positive skewness, the extreme data values are greater, which in turn
increases the mean value of the dataset.
In positive skewness: Mode < Median < Mean.

Negative Skewness: In negative skewness, the extreme data values are smaller, which in turn
decreases the mean value of the dataset.

In negative skewness: Mean < Median < Mode.

Note: When there is no skewness in the dataset, it is referred to as Symmetrical Distribution.


In Symmetrical Distribution: Mean = Median = Mode.

DEPARTMENT OF COMPUTER APPLICATION Page | 3


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Fourth Moment – Kurtosis


Kurtosis is the fourth moment, which measures the presence of outliers in the distribution i.e.,
it gives the graph of either heavily-tailed or lightly-tailed due to the presence of outliers. In
simple terms, kurtosis measures the peakedness or flatness of a distribution.
 If the graph has a shorter tail and a flat top, then Kurtosis is said to be high.
 If the graph has a higher peak and lower tail, then the kurtosis is said to be low.
∑(𝐱ᵢ – µ)𝟒
µ4 = 𝒏
∑(𝐱ᵢ – µ)𝟒 µ𝟒
Kurtosis (β) = =
𝒏∗µ𝟐 𝟐 µ𝟐 𝟐

There are three types of Kurtosis:


Mesokurtic: This is the same as Normal distribution, i.e., a type of distribution where both the
extreme ends of the graph are similar.

Lepotokurtic: This distribution indicates that a more significant percentage of data is present
near the tail, which implies the longer tail. Lepotokurtic has a greater value of kurtosis than
Mesokurtic.

Platykurtic: This distribution indicates that there is fewer data in the tail portion, which
implies a shorter tail. Platykurtic has a lesser value of kurtosis than Mesokurtic.

Interpretation of moment statistics


Mean – 1st moment about the origin – central tendency measure
Variance- 2nd moment about mean – dispersion measure
Skewness – 3rd standardize moment – skewness measure
DEPARTMENT OF COMPUTER APPLICATION Page | 4
Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

µ3 = 0 → symmetrical
µ3 > 0 → positively skewed
µ3 < 0 → negatively skewed
For µ3 between ± 0.2, the distribution can be assumed to be normal with respect to skewness.

Kurtosis – 4th moment- kurtosis measure


µ4 − 3 = 0 → same peakedness as normal curve.
µ4 − 3 > 0 → more peakedness as normal curve.
µ4 − 3 = 0 → flatter than normal curve.
For µ4 – 3 between ± 0.5, the curve can b consider normal with respect to kurtosis.

Mathematical Expectation
Mathematical expectation, also known as the expected value, which is the summation of all
possible values from a random variable.
It is also known as the product of the probability of an event occurring, denoted by P(x), and the
value corresponding with the actually observed occurrence of the event.
For a random variable expected value is a useful property. E(X) is the expected value and can be
computed by the summation of the overall distinct values that is the random variable. The
mathematical expectation is denoted by the formula:
E(X)= Σ (x1p1, x2p2, …, xnpn),

DEPARTMENT OF COMPUTER APPLICATION Page | 5


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Where, x is a random variable with the probability function, f(x),


p is the probability of the occurrence, and n is the number of all possible values.
The mathematical expectation of an indicator variable can be 0 if there is no occurrence of an
event A, and the mathematical expectation of an indicator variable can be 1 if there is an
occurrence of an event A.
For example, a dice is thrown, the set of possible outcomes is {1,2,3,4,5,6} and each of this
outcome has the same probability 1/6. Thus, the expected value of the experiment will be
1/6*(1+2+3+4+5+6) = 21/6 = 3.5. It is important to know that “expected value” is not the same
as “most probable value” and, it is not necessary that it will be one of the probable values.

Properties of Expectation
1. If X and Y are the two variables, then the mathematical expectation of the sum of the two
variables is equal to the sum of the mathematical expectation of X and the mathematical
expectation of Y.
Or
E(X + Y) = E(X) + E(Y)
2. The mathematical expectation of the product of the two random variables will be the
product of the mathematical expectation of those two variables, but the condition is that the
two variables are independent in nature. In other words, the mathematical expectation of
the product of the n number of independent random variables is equal to the product of the
mathematical expectation of the n independent random variables
Or
E(XY)=E(X)E(Y)
3. The mathematical expectation of the sum of a constant and the function of a random
variable is equal to the sum of the constant and the mathematical expectation of the function
of that random variable.
Or
E (a + f(X)) = a + E (f(X)),
Where, a is a constant and f(X) is the function.

4. The mathematical expectation of the sum of product between a constant and function of a
random variable and the other constant is equal to the sum of the product of the constant
and the mathematical expectation of the function of that random variable and the other
constant.
Or
E (a X + b) = a E(X) + b

5. The mathematical expectation of a linear combination of the random variables and constant
is equal to the sum of the product of ‘n’ constant and the mathematical expectation of the
‘n’ number of variables.
Or
E(∑ai Xi)=∑ ai E(Xi)
Where, ai, (i=1…n) are constants.

DEPARTMENT OF COMPUTER APPLICATION Page | 6


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Dispersion
The field of statistics is used across every sector and industry to help people better understand,
and predict, potential outcomes. In finance, investors often turn to statistics to gain a sense of
how returns on certain assets, or groups of assets, could be distributed. This range of possible
investment returns is called dispersion. In other words, dispersion refers to the range of
potential outcomes of investments based on historical volatility or returns.
There are two important ways to measure dispersion—alpha and beta—which calculate risk-
adjusted returns and returns relative to a benchmark, respectively. By considering the
dispersion of possible investment returns and values such as alpha and beta, investors can gain
a sense of the risk inherent in a particular security or investment portfolio.

Measures of Dispersion
In statistics, the measures of dispersion help to interpret the variability of data i.e. to know how
much homogenous or heterogeneous the data is. In simple terms, it shows how squeezed or
scattered the variable is.
Types of Measures of Dispersion
There are two main types of dispersion methods in statistics which are:

 Absolute Measure of Dispersion


 Relative Measure of Dispersion

Absolute Measure of Dispersion


An absolute measure of dispersion contains the same unit as the original data set. The absolute
dispersion method expresses the variations in terms of the average of deviations of observations
like standard or means deviations. It includes range, standard deviation, quartile deviation, etc.
The types of absolute measures of dispersion are:

1. Range: It is simply the difference between the maximum value and the minimum value
given in a data set. Example: 1, 3,5, 6, 7 => Range = 7 -1= 6
2. Variance: Deduct the mean from each data in the set, square each of them and add each
square and finally divide them by the total no of values in the data set to get the variance.
Variance (σ2) = ∑(X−μ)2/N
3. Standard Deviation: The square root of the variance is known as the standard
deviation i.e. S.D. =√𝜎.
4. Quartiles and Quartile Deviation: The quartiles are values that divide a list of
numbers into quarters. The quartile deviation is half of the distance between the third
and the first quartile.
5. Mean and Mean Deviation: The average of numbers is known as the mean and the
arithmetic mean of the absolute deviations of the observations from a measure of central
tendency is known as the mean deviation (also called mean absolute deviation).

DEPARTMENT OF COMPUTER APPLICATION Page | 7


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Quartiles and Quartile Deviation:-


𝑵
𝒓( )−𝒄
𝟒
Qr = l1 + (l2 − l1)
𝒇
Qr = the rth quartile
l1 = the lower limit of the quartile class
l2 = the upper limit of the quartile class
f = the frequency of the quartile class
c = the cumulative frequency of the class preceding the quartile class
N = Number of observations in the given data set

Relative Measure of Dispersion


The relative measures of dispersion are used to compare the distribution of two or more data
sets. This measure compares values without units. Common relative dispersion methods
include:

1. Co-efficient of Range
2. Co-efficient of Variation
3. Co-efficient of Standard Deviation
4. Co-efficient of Quartile Deviation
5. Co-efficient of Mean Deviation

Co-efficient of Dispersion
The coefficients of dispersion are calculated (along with the measure of dispersion) when two
series are compared, that differ widely in their averages. The dispersion coefficient is also used
when two series with different measurement units are compared. It is denoted as C.D.
The common coefficients of dispersion are:
C.D. in terms of Coefficient of dispersion
Range C.D. = (Xmax – Xmin) ⁄ (Xmax + Xmin)
Quartile Deviation C.D. = (Q3 – Q1) ⁄ (Q3 + Q1)
Standard Deviation (S.D.) C.D. = S.D. ⁄ Mean
Mean Deviation C.D. = Mean deviation/Average

Dispersion Measures: Characteristics


 It must be simple to calculate and comprehend.
 The fluctuations in observations haven’t had much of an impact.
 Based on everything you’ve seen
 It should be able to be subjected to further mathematical analysis.
 It should be influenced by sampling fluctuations.
DEPARTMENT OF COMPUTER APPLICATION Page | 8
Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Expectation of Two-Dimensional Random Variables


Introduction

In the previous chapter we studied various aspects of the theory of a single R.V. In this chapter
we extend our theory to include two R.V's one for each coordinator axis X and Y of the XY
Plane.
Definition: Let S be the sample space. Let X = X(S) & Y = Y(S) be two functions
each assigning a real number to each outcome s ∈ S. hen (X, Y) is a two dimensional random
variable.

Types of random variables


1. Discrete R.V.’s
2. Continuous R.V.’s

Discrete R.V.’s (Two Dimensional Discrete R.V.’s)


If the possible values of (X, Y) are finite, then (X, Y) is called a two dimensional discrete R.V.
and it can be represented by (xi, y), i = 1,2,….,m.
In the study of two dimensional discrete R.V.’s we have the following 5 important terms.

 Joint Probability Function (JPF) (or) Joint Probability Mass Function.


 Joint Probability Distribution.
 Marginal Probability Function of X.
 Marginal Probability Function of Y.
 Conditional Probability Function.

1.1 Joint Probability Function of discrete R.V.’s X and Y


The function P(X = xi, Y = yj) = P(xi, yj) is called the joint probability function for
discrete random variable X and Y is denote by pij.
Note:-
1. P(X = xi, Y = yj) = P[(X = xi)∩(Y = yj)] = pij
2. It should satisfies the following conditions
(i) pij ≥ ∀ i, j (ii) ΣjΣi pij = 1
1.2

1.2 Marginal Probability Function of X


If the joint probability distribution of two random variables X and Y is given then the marginal
probability function of X is given by
Px(xi) = pi
Marginal PMFs of X and Y
:
DEPARTMENT OF COMPUTER APPLICATION Page | 9
Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

2 CONTINUOUS RANDOM VARIABLES

• Two dimensional continuous R.V.’s


If (X, Y) can take all the values in a region R in the XY plans then (X, Y) is called two-
dimensional continuous random variable.

 Joint density function


1. fXY(x,y) ≥ 0 ;
𝑥 𝑦
2. ∫−∞ ∫−∞ 𝑓𝑥𝑦 (x, y) 𝑑𝑥𝑑𝑦 = 1
 Joint probability distribution function
1. Fxy = P[X ≤ x, Y ≤ y]
𝑥 𝑦
= ∫−∞{∫−∞ 𝑓𝑥𝑦 (x, y) 𝑑𝑥}𝑑𝑦

 Marginal Probability density function



f(x) = fx(x) = ∫−∞ 𝑓𝑥,𝑦 (x, y) 𝑑𝑦 ( marginal pdf of X)


f(y) = fy(y) = ∫−∞ 𝑓𝑥,𝑦 (x, y) 𝑑𝑥 ( marginal pdf of Y)

 Conditional probability density function


𝑓(𝑥,𝑦)
1. P(Y=y / X=x ) = f (y/x) = 𝑓(𝑥)
, f(x) >0
𝑓(𝑥,𝑦)
2. P(X=x / Y=y ) = f (x / y) = 𝑓(𝑦) , f(y) >0

REGRESSION
Line of regression
The line of regression of X on Y is given by
𝜎𝑦
𝑥 − 𝑥 = r.𝜎 (𝑦 − 𝑦̅)
𝑥

The line of regression of Y on X is given by


𝜎𝑦
𝑦 − 𝑦̅ = r.𝜎 (𝑥 − 𝑥 )
𝑥

DEPARTMENT OF COMPUTER APPLICATION Page | 10


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Angle between two lines of regression


1 − 𝑟2 𝜎𝑦 𝜎𝑥
tan 𝜃 = ( )
𝑟 𝜎𝑥 2 + 𝜎𝑦 2

Regression coefficient

 Regression coefficient of Y on X
𝜎𝑦
r.𝜎 = bYX
𝑥

 Regression coefficient of X on Y
𝜎
r.𝜎𝑥 = bXY
𝑦

 Correlation coefficient r = ±√𝑏𝑥𝑦 𝑏𝑦𝑥


∑(𝑥𝑖 −𝑥)(𝑦𝑖 −𝑦̅)
Cov[X,Y] = 𝑁−1

Linear correlation
Linear correlation is a measure of dependence between two random variables.
It has the following characteristics:
 it ranges between -1 and 1;
 it is proportional to covariance;
 its interpretation is very similar to that of covariance .
Correlation
 Let x and y be two random variables
 The linear correlation coefficient between x and y is
𝐶𝑜𝑣[𝑋,𝑌]
 Cor[X,Y] =
𝜎𝑋 𝜎𝑌
 Where: Cov[X,Y] is the covariance between x and y.

DEPARTMENT OF COMPUTER APPLICATION Page | 11


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Zero standard deviations


In principle, the ratio is well defined only if std[X] and std[y] are strictly greater than zero.
However, it is often assumed that Cov[X,Y]=0 when one of the two standard deviation is
zero.
This is equivalent to assuming that 0/0 = 0 because Cov[X,Y] = 0 when one of the two
standard.

Interpretation
The interpretation is similar to the interpretation of covariance: the correlation between X and
Y provides a measure of how similar deviation from the respective means are.

DEPARTMENT OF COMPUTER APPLICATION Page | 12


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Linear correlation ranges between -1 to 1.


-1 ≤ Cov[X,Y] ≤ 1

Thanks to this property, correlation allows us to easily understand the intensity of the linear
dependence between two random variables.
 The clser correlation is to 1, the stronger the positive linear dependence between X
and Y is.
 The closer it is to -1, the stronger the negative linear dependence between X and Y is.
Terminology
The following terminology is often used
1. If Cov[X,Y] > 0, then X and Y are said to be Positively linearly correlated.
2. If Cov[X,Y] < 0, then X and Y are said to be Negatively linearly correlated.
3. If Cov[X,Y] = 0, then X and Y are said to be uncorrelated.
4. If Cov[X,Y] ≠ 0, then X and Y are said to be linearly correlated.
Sample in Statistics: An Introduction to Sampling
Sampling is very often used in our daily life. For example, while purchasing fruits from a shop,
we usually examine a few to assess the quality. A doctor examines a few drops of blood as a
sample and draws a conclusion about the blood constitution of the whole body.

Sample in Statistics: The Process of Sampling


Sampling is the process of selecting units (e.g., people, organizations) from a population of
interest so that by studying the sample we may fairly generalize our results back to the
population from which they were chosen. Such is a sample in statistics.
The sampling of a sample in statistics works in the following manner:

1. The first step is to define the population of interest


2. The second step is to specify the sampling frame. The sampling frame is a set of events
which are possibly measurable. In other words, it refers to the source material or device
from which a sample is taken; it may include individuals, households, and institutions
too
3. The third step is to specify a sampling method
4. The fourth step is to determine the sample size
5. The fifth step is to implement the sampling plan or execute the sampling plan
6. The sixth step is, to begin with the actual sampling and data collection
Taking a sample in statistics is an efficient and effective alternative to looking at all the data.
Sampling a sample in statistics helps you to:

 Collect a portion of all the data


 Use that portion of the data to make conclusions
 Save on time, resources, and money

DEPARTMENT OF COMPUTER APPLICATION Page | 13


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Sample in Statistics: Sampling Error


As with anything in life, there is also room for error in the sampling of a sample in
statistics. If you would analyze the whole population, there would be no sampling error. The
first type of sampling error comes from the fact that the parameter that you measure in the
sample would not be 100% the same as for the entire population. Furthermore, if you take
a sample three times, the parameter would also be slightly different between the three samples.
We call this the margin of error. This margin of error is unavoidable in collecting a sample
in statistics.
The second type of error that can occur during sampling samples in statistics is the non-
sampling error. These errors occur when the method of sampling is faulty. It can be due to
faulty measurements or due to incorrect sampling methods that caused bias or due to
mistakes in the data. These types of errors for collecting a sample in statistics will provide data
that can’t be trusted. This is why it is very important that the sampling method that you use is
trustworthy and that that the measuring tool that you use is trustworthy as well. This type of
error can easily be avoided by following the correct sampling principles for obtaining a correct
sample in statistics.
A sample in statistics is basically the group of subjects that Six Sigma teams will analyze to
make assumptions about the population that they want to study. It is usually impossible to
collect data from an entire population, which is why sampling is so important. There are many
different sampling techniques, but these can be categorized broadly as probability sampling
and non-probability sampling. But more on that later. The most important thing to know is that
correct sampling procedures must be used to obtain a representative sample in statistics.

T Tests
A t test is a statistical test that is used to compare the means of two groups. It is often used
in hypothesis testing to determine whether a process or treatment actually has an effect on the
population of interest, or whether two groups are different from one another. When to use
a t test
A t test can only be used when comparing the means of two groups
The t test is a parametric test of difference, meaning that it makes the same assumptions about
your data as other parametric tests.

The t test assumes your data:


1. are independent
2. are (approximately) normally distributed
3. have a similar amount of variance within each group being compared (a.k.a.
homogeneity of variance)
If your data do not fit these assumptions, you can try a nonparametric alternative to the t test,
such as the Wilcoxon Signed-Rank test for data with unequal variances.

T test formula
One-Sample T-Test
While performing this test, the mean or average of one group is compared against the set
average, which is either the theoretical value or means of the population. For example, a teacher
wishes to figure out the average height of the students of class 5 and compare the same against
a set value of more than 45 kgs.

DEPARTMENT OF COMPUTER APPLICATION Page | 14


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

The teacher first randomly selects a group of students and records individual weights to achieve
this. Next, she finds out the mean weight for that group and checks if it meets the standard set
value of 45+. The formula used to obtain one-sample t-test results is:

Where,

 t = t-statistic
 m = mean of the group
 𝜇 = theoretical mean value of the population
 s = standard deviation of the group
 n = sample size
Independent Two-Sample T-Test
This is the test conducted when samples from two different groups, species, or populations are
studied and compared. It is also known as an independent T-test. For example, if a teacher
wants to compare the height of male students and female students in class 5, she would use the
independent two-sample test.
The T-test formula used to calculate this is:

Where,

 mA – mB = means of samples from two different groups or populations


 nA – nB = respective sample sizes
 s2 = standard deviation or common variance of two samples

Paired Sample T-Test


This hypothesis testing is conducted when two groups belong to the same population or group.
The groups are studied either at two different times or under two varied conditions. The formula
used to obtain the t-value is:

Where,
 T = t-statistic
 m = mean of the group
 𝜇 = theoretical mean value of the population
 s = standard deviation of the group
 n = sample size

DEPARTMENT OF COMPUTER APPLICATION Page | 15


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Equal Variance T-Test


This test is conducted when the sample size in each group or population is the same or the
variance of the two data sets is similar. It is also referred to as pooled T-test. The
formula applied here is as follows:

Where,

 Mean1 and mean2 = average value of each set of samples


 var1 and var2 = variance of each set of samples
 n1 and n2 = number of records in each set

Unequal Variance T-Test


The unequal variance testing is used when the variance and the number of samples in each
group are different. It is often referred to as Welch’s test, and the formula is:

Where,

 mean1 and mean2 = Average value of each set of samples


 var1 and var2 = Variance of each set of samples
 n1 and n2 = number of records in each set

One-sample t-test Two-sample t-test Paired t-test


 Independent groups t-test  Paired groups t-test
Synonyms Student’s t-test
 Independent samples t-test  Dependent samples t-test
 Equal variances t-test
 Pooled t-test
 Unequal variances t-test
Number of
variables One Two Two
Type of
variable Continuous Continuous measurement Continuous measurement
measurement
Categorical or Nominal to Categorical or Nominal to
define groups define pairing within group

DEPARTMENT OF COMPUTER APPLICATION Page | 16


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Purpose of
test Decide if the Decide if the population Decide if the difference
population mean means for two different between paired
is equal to a groups are equal or not measurements for a
specific value or population is zero or not
not
Example:
test if... Mean heart rate Mean heart rates for two Mean difference in heart
of a group of groups of people are the rate for a group of people
people is equal to same or not before and after exercise is
65 or not zero or not
Estimate of
population Sample average Sample average for each Sample average of the
mean group differences in paired
measurements
Population
standard Unknown, use Unknown, use sample Unknown, use sample
deviation sample standard standard deviations for each standard deviation of
deviation group differences in paired
measurements
Degrees of
freedom Number of Sum of observations in each Number of paired
observations in sample minus 2, or: observations in sample
sample minus 1, n1 + n2 – 2 minus 1, or:
or: n–1
n–1

F-Test Definition
F-Test in statistics is a hypothesis-testing procedure that considers two variances from two
samples. The F-Test is used when the difference between two variances needs to be
significantly assessed, i.e., when determining whether or not two samples can be taken as
representative of the normal population with the same variance.

The F-Test helps to determine the overall significance of the regression. It is useful in various
situations, such as when a quality controller wants to determine whether the product’s quality
is deteriorating over time. In addition, it might be useful for an economist to determine whether
income variability varies between two populations.

Key Takeaways

 The F-test is a statistical test that evaluates if the variances of the two normal populations
are equal.
 One can deem the variance ratio of the test insignificant if F O= 0.5, and one can assume
that the values will be from the same group or groups with similar variances.
 The null hypothesis is rejected, and the variance ratio is considered significant if F≥0.5.

DEPARTMENT OF COMPUTER APPLICATION Page | 17


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

 The F-test vs. t-test: The t-test and the F-test are two separate tests. The T-test compares two
populations’ means, whereas the other compares two populations’ variances.

F-Test in Statistics Explained


F-test in statistics helps to decide whether two populations’ variances are equal. This is the
variance ratio test because it calculates the ratio of variances. The goal of the test is to
determine whether the variance in two populations is equal. It was propounded by British
polymath R.A. Fisher and named to honor him. G.W. Snedecor later developed the test.
The following conditions are critical for using the F-test to compare the variances of two
populations:

1. Normality: the populations must have a normal distribution.


2. Independent and random selection of sample items: the selection of the samples’
components should be independent and random.
3. More than unity: The variance ratio must be one or larger than one; it cannot be less than
one. When dividing variance estimates, smaller estimates divide the larger estimates of
variances.
4. The additive property states that the total of different variance components will equal the
total variance, i.e., the total variance is the sum of the variance between samples and the
variance within samples.
Formula
1. Sample variances: The formula for calculating sample variances is as follows (an
online F-test calculator can make it easier):

∑(𝑥1 −𝑥 )2 ∑(𝑥2 −𝑥)2


S12 (sample1) = And Variance sample for S22(sample 2) =
𝑛1 −1 𝑛2 −1

2. Null hypothesis: After the formation of the test, the null hypothesis are either
(a) Two samples were from the same group or
(b) The population’s variances concerning both samples are equal.
3. To compute the variance ratio, use the formula F = larger estimate divided by a smaller
estimate of variance. Regardless of whether S12 or S22, the numerator will always be the
larger value.
4. When calculating degrees of freedom, the larger the sample’s variance is V1; the smaller
variance is V2.
5. Table value of F: the critical value of F is available from the “F-Table” (F-test table) at
the determined significance level.
6. Analysis: This involves the comparison of the computed value and the tabulated value. For
various levels of significance, there are several F Tables (F-test tables).
(a) The variance ratio is insignificant if F ≤ F0.5. We can assume that the values are from the
same group or groups with similar variances.

(b) The null hypothesis is rejected, and the variance ratio is considered significant if F ≥ 0.5.
Consider the example of the population in a village:

DEPARTMENT OF COMPUTER APPLICATION Page | 18


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Village A B
Sample size 10 12
Mean monthly income 150 140
Sample variance 92 110

Testing the equality of sample variances with a significance level of 5% with the above-given
date.
Null Hypothesis: - Ho :- 𝜎12 = 𝜎22
∑(𝑥1 −𝑥)2 92
Variance sample for S12 (sample1) = = 10−1 = 10.22
𝑛1 −1
And
∑(𝑥2 −𝑥)2 110
Variance sample for S22(sample 2)= = 12−1 = 10
𝑛2 −1

𝑆12 10.22
F= = = 1.022
𝑆22 10
The critical value for v1 (10-1) = 9 and v2 (12-1) =11 and the table value of F at 5%
significance level = 2.90. An online f-test calculator can help you in making the calculations
easier.
Interpretation
The F statistic helps to decide whether to accept or reject the null hypothesis. The test results
must include an F value and an F critical value. The F value is compared to a particular value
known as the F critical value. The value derived from the data is the F statistic, or F value
(without the “critical” part). In general, one can reject the null hypothesis if the computed F
value for a test is higher than the F critical value.
In the example above, F’s computed value (1.022) is less than its table value obtained from
the F table (2.90). As a result, one can conclude that the null hypothesis is true and that the
variance of the two samples is equal.
F-Test for Equality of Two Variances

DEPARTMENT OF COMPUTER APPLICATION Page | 19


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

F - Test T - Test
An F test is a test statistic used to check the The T-Test is used when the sample size is
equality of variance of two population small ( n < 30) and the population standard
deviation is not known.

The data follows an F distribution The data follow T distribution.


𝑆12 The T-Test statistics for 1 sample is given
The F test statistic is given as F = 𝑆22 𝑥−𝜇
by T = 𝑠 , where 𝑥 𝑖𝑠 the sample mean ,
√𝑛

The F test is used for variance This is used for testing means

Chi-Square Test for Goodness of Fit


Goodness-of-Fit Test

In this type of hypothesis test, you determine whether the data “fit” a particular distribution or
not. For example, you may suspect your unknown data fit a binomial distribution. You use a
chi-square test (meaning the distribution for the hypothesis test is chi-square) to determine if
there is a fit or not. The null and the alternative hypotheses for this test may be written in
sentences or may be stated as equations or inequalities.
∑(𝑂−𝐸)2
The test statistics for goodness of fit test is :- ∑𝐸
Where: -
O = Observed Value
E = expected value

Definition of Attributes:
The dictionary meaning of attribute is quality or characteristic.
Examples: Attributes are beauty, honesty, gender (male & female), health, employment,
smoking, drinking, blindness, etc.
The theory of attributes deals with qualitative characteristics which cannot be measured
quantitatively. In the study of attributes, the objects (units of the population/sample) are
classified according to the presence or absence of the attribute (quality) in them.
Example: A person may be classified as smoker or non-smoker, blind or not blind, male or
female, employed or un-employed, healthy or sick, literate of illiterate and so on.

Types of Attributes
Categorical Attributes
These attributes represent qualitative or nominal data that cannot be measured or ordered, and
they can be further classified into:
 Binary attributes: These attributes can take only two values, such as yes or no, true or
false, male or female.
 Nominal attributes: These attributes have multiple categories that do not have any
natural order, such as colors, shapes, or names of people.

DEPARTMENT OF COMPUTER APPLICATION Page | 20


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Numerical Attributes
These attributes represent quantitative data that can be measured or ordered, and they can be
further classified into:
 Discrete attributes: These attributes take on integer values, such as the number of
siblings, or the number of cars in a parking lot.
 Continuous attributes: These attributes can take on any real value within a range, such
as weight, height, or temperature.
Ordinal Attributes
These attributes represent data that can be ordered or ranked, but the differences between the
values may not be uniform, and they can be further classified into:
 Interval attributes: These attributes have a fixed scale with uniform intervals between
values, such as dates, time, or temperature measured in Celsius or Fahrenheit.
 Ratio attributes: These attributes have a fixed scale with a true zero point, such as weight,
height, or distance.
Textual Attributes
These attributes represent data in natural language format, such as text documents, emails, or
social media posts.
Geographic Attributes
These attributes represent data related to geographic locations, such as latitude, longitude, or
postal codes.
Temporal Attributes
These attributes represent data related to time, such as dates, times, or durations.
 Date attributes: These attributes represent a specific date, such as birthdate or expiration
date.
 Time attributes: These attributes represent a specific time of day, such as the time an
event starts or ends.
 Duration attributes: These attributes represent a length of time, such as the duration of a
movie or the time it takes to complete a task.
Boolean Attributes
These attributes represent data that can take only two possible values, such as true or false, yes
or no, or on or off.
Image Attributes
These attributes represent data related to images, such as size, resolution, or color depth.
Audio Attributes
These attributes represent data related to audio signals, such as sampling rate, bit depth, or
duration.
Video Attributes
These attributes represent data related to video signals, such as frame rate, resolution, or
codec.
Structural Attributes
These attributes represent data related to the structure or hierarchy of objects, such as parent-
child relationships or the depth of a node in a tree.
Behavioral Attributes
These attributes represent data related to the behavior or actions of objects, such as frequency
of use or the number of clicks on a website.

DEPARTMENT OF COMPUTER APPLICATION Page | 21


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Examples of Attribute
Here are some examples of attributes:
 Color: The color of an object is an attribute. For example, a red apple or a blue car.
 Size: The size of an object is an attribute. For example, a large house or a small dog.
 Shape: The shape of an object is an attribute. For example, a round ball or a square box.
 Texture: The texture of an object is an attribute. For example, a smooth surface or a rough
texture.
 Material: The material of an object is an attribute. For example, a wooden chair or a metal
spoon.
 Age: The age of a person or an object is an attribute. For example, a young child or an
antique vase.
 Gender: The gender of a person or an animal is an attribute. For example, a male cat or a
female bird.
 Height: The height of a person or an object is an attribute. For example, a tall building or
a short person.
 Weight: The weight of a person or an object is an attribute. For example, a heavy rock or
a lightweight backpack.
 Personality: The personality of a person is an attribute. For example, a kind-hearted
person or a funny friend.

Applications of Attribute
Here are some common applications of attributes in various fields:
 Descriptive statistics: Attributes can be used to describe the characteristics of a sample
or population in a research study. For example, attributes such as age, gender, income, and
education level can be used to describe the demographic characteristics of participants in
a study.
 Sampling: Attributes can be used to define the population of interest and to select a sample
from that population. For example, attributes such as geographic location, age, or income
can be used to define a target population for a study, and then used to select a sample that
represents that population.
 Programming: In programming, attributes are used to define properties of objects or
classes. Attributes can be used to specify metadata, such as the author or version of code,
or to provide additional information about the behavior of a method or class.
 Data analysis: Attributes are used to describe the characteristics of data sets, such as their
size, shape, and distribution. Attributes can also be used to indicate missing or incomplete
data points.
 Web development: HTML attributes are used to specify the properties of elements on a
web page, such as the color of text, the size of an image, or the target of a link.
 Machine learning: In machine learning, attributes are used to describe the features of
input data, such as the height and weight of a person in a dataset about human bodies.
 Database management: Attributes are used to define the fields in a database table, such
as the name and age of a person in a database about customers.
 Graphic design: Attributes such as color, shape, and texture are used to create visual
elements in graphic design, such as logos, icons, and user interface elements.

DEPARTMENT OF COMPUTER APPLICATION Page | 22


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

Purpose of Attribute
The purpose of an attribute is to provide information about the characteristics or qualities of an
object, person, or concept. Attributes can help us to identify and distinguish things from one
another, and they can also help us to make decisions and draw conclusions based on the
information provided.
Attributes are important in many areas of life, including:
 Research: Attributes can be used to collect data and identify patterns, which can help
researchers to draw conclusions and make predictions.
 Product design: The attributes of a product, such as its size, color, and material, can help
designers to create products that meet the needs and preferences of their target audience.
 Marketing: Attributes can be used to promote and differentiate products from their
competitors, by highlighting the unique features and benefits.
 Personal development: Understanding our own attributes can help us to identify our
strengths and weaknesses, and work on improving ourselves.

Characteristics of Attribute
Some of the key characteristics of attributes are:
 Identifiability: An attribute must be distinguishable and identifiable from other
characteristics or qualities of an object or person.
 Descriptiveness: Attributes should be descriptive, meaning they provide information
about the quality or characteristic of an object or person.
 Measurability: Attributes can be measured or quantified in some way. For example, the
weight of an object or the age of a person.
 Relevance: Attributes should be relevant to the context or situation in which they are being
used. For example, the color of a car is not relevant when discussing the car’s fuel
efficiency.
 Consistency: Attributes should be consistent over time and across different contexts. For
example, if an object is red, it should be consistently red in all lighting conditions.
 Importance: Attributes should be important or meaningful in the context they are being
used. For example, the material of a product may be more important than its color when
making a purchasing decision.
 Objectivity: Attributes should be objective, meaning they are based on facts and not
influenced by personal bias or opinion.

Advantages of Attribute
Attributes have several advantages, including:
 Clarity and Precision: Attributes help to provide a clear and precise description of an
object, person, or concept. They allow us to break down complex ideas or characteristics
into smaller, more manageable parts, making it easier to understand and communicate
them.
 Comparability: Attributes allow for easy comparison between different objects, people,
or concepts. By identifying and measuring similar attributes, we can compare and evaluate
them based on their relative strengths and weaknesses.
 Objectivity: Attributes are objective, meaning they are based on facts and not influenced
by personal bias or opinion. This makes them a useful tool for analyzing and describing
objects or concepts in a consistent and unbiased manner.

DEPARTMENT OF COMPUTER APPLICATION Page | 23


Proprietary material of SILVER OAK UNIVERSITY
SUBJECT
Statistical Method &
Probability(3040003102)
SEMESTER: 2 PREPARED BY : Prof. Keyuri Shah

 Efficiency: Attributes provide a quick and efficient way to collect and organize
information about an object or person. This can save time and resources when making
decisions or conducting research.
 Customization: Attributes can be customized to fit specific needs or contexts. For
example, attributes can be tailored to meet the unique requirements of a particular product
or market segment.
 Predictive Power: Attributes can help to predict the behavior or performance of an object
or person. By identifying key attributes that are associated with success or failure, we can
make more informed decisions and take actions to improve outcomes.

Limitations of Attribute
Some Limitations of Attribute are as follows:
 Limited Scope: Attributes can only provide a limited description of an object, person, or
concept. They may not capture all of the nuances and complexities that make up the whole.
 Context Dependency: Attributes may be dependent on the context or situation in which
they are being used. For example, the color of a product may be less important than its
durability in certain contexts.
 Subjectivity: Despite efforts to make attributes objective, there is always the potential for
personal bias or subjectivity to influence the selection and interpretation of attributes.
 Over-Simplification: Attributes can sometimes over-simplify the characteristics of an
object, person, or concept, leading to an incomplete or inaccurate understanding.
 Incomplete Coverage: Attributes may not cover all of the important characteristics or
qualities of an object, person, or concept. This can lead to an incomplete or inaccurate
understanding of the whole.
 Lack of Flexibility: Attributes can be inflexible and may not adapt well to changing
circumstances or new information.
 Difficulty of Measurement: Some attributes may be difficult to measure or quantify,
making them less useful for analysis or comparison.

DEPARTMENT OF COMPUTER APPLICATION Page | 24


Proprietary material of SILVER OAK UNIVERSITY

You might also like