Block 1
Block 1
STATISTICAL
Indira Gandhi National Open University
School of Sciences INFERENCE
Sampling Sampling
distribution distribution
under under
H0 H1
ᵝ α
H0: θ = θ0 H1: θ = θ1>θ0
ndom Selectio
Ra n
Sample
Population
Sampling distribution
of test statistic
R2
(1- α) Non-rejection
region
Rejection/ Rejection/
Critical region R3 R5 Critical region
R
α/2 6 α/2 R9
R4
Volume
1
FOUNDATION OF STATISTICAL INFERENCE
BLOCK 1
Sampling Distributions 5
BLOCK 2
Properties of Good Estimator 159
Appendix 243
Curriculum and Course Design Committee
Prof. Sujatha Varma Prof. Rakesh Srivastava
Former Director, SOS Department of Statistics
IGNOU, New Delhi M.S. University of Baroda, Vadodara (GUJ)
Print Production
Mr. Tilak Raj
Assistant Registrar
MPDD, IGNOU, New Delhi
November, 2024
© Indira Gandhi National Open University, 2024
ISBN-978-81-266-
All rights reserved. No part of this work may be reproduced in any form, by mimeograph or any other means,
without permission in writing from the Indira Gandhi National Open University
Further information on the Indira Gandhi National Open University may be obtained from the University’s Office at
Maidan Garhi, New Delhi-110068 or visit University’s website http://www.ignou.ac.in
Printed and published on behalf of the Indira Gandhi National Open University, New Delhi by the Director, School
of Sciences.
Printed at: Raj Printers, A-9, Sector B-2, Tronica City Loni, Ghaziabad, UP-201102
VOLUME 1: FOUNDATION OF STATISTICAL
INFERENCE
Dear learners, welcome to the course titled, “Statistical Inference”. In many situations, we
have to extract some information from all elements or units of a large population, in general, it
may be time-consuming and expensive. Also, if the elements or items of a population are
destroyed under investigation then gathering information from all elements of the population is
not making sense because all elements will be destroyed under investigation. Therefore, in
most of the cases in daily life, business and industry, the information about the population is
gathered through a random sample. The results of a properly taken sample enable the
investigator to arrive at generalisations that are valid for the entire population. The process of
generalising sample results to the population is called Statistical Inference. The statistical
inference may be divided into two categories:
(i) If the population characteristics (known as parameters) are unknown and we guess the
true value of the unknown parameters on the basis of a random sample drawn from the
population, then this technique is known as “Estimation”.
(ii) If some information is available about the population characteristics and we verify whether
the information is true or not on the basis of a random sample drawn from the population,
then this technique is known as “Testing of hypothesis”.
The aim of this course is to develop the skills of learners to estimate and testing of hypothesis
about the population characteristics such as means, proportions, variances, etc.
This theory course is worth four credits and is comprised of eighteen units that are organised
into four blocks into two volumes each of two credits which are given as follows:
Volume 1: Foundation of Statistical Inference
Volume 2: Estimation and Parametric Tests
Volume 1 contains the first two blocks of the course.
The first block in this volume is titled “Sampling Distributions” and it contains five units that
broadly cover the various sampling distributions of commonly used statistics such as mean,
proportion, variance, difference of two means, difference of two proportions and ratio of two
variances. The most important theorem “central limit theorem” and law “law of large numbers”
of Statistics are also described with their applications. This block describes some important
sampling distributions which are widely used in statistical inference and known as exact
sampling distributions such as chi-square, t and F and how to read their tables.
The second block of this volume is titled “Properties of Good Estimator” and it contains four
units that broadly cover various properties of a good estimator. In this block, we shall discuss
and explain the concepts of unbiasedness, consistency, asymptotically normal consistency,
efficiency, most efficient estimator, mean squared error, minimum variance unbiased
estimator, sufficiency and minimal sufficient estimator.
If you feel like reading more than what this course contains, you may like to consult the
following books:
Feedback Link:
https://docs.google.com/forms/d/e/1FAIpQLSdfldMvjNLVADrE49VleFxpX9paYwS-
0_UVkX3crO2z3G9DrA/viewform?usp=sf_link
We hope that you would enjoy reading the self-learning material of this course. Wishing you a
happy learning experience and all the best in this endeavour!
Block
1
SAMPLING DISTRIBUTIONS
UNIT 1
Basic Concepts of Sampling Distributions 9
UNIT 2
Sampling Distributions of Sample Means 47
UNIT 3
Sampling Distributions of Sample Proportions and Variances 81
UNIT 4
Sampling Distributions Associated with Normal Populations-I 109
UNIT 5
Sampling Distributions Associated with Normal Populations-II 131
5
BLOCK 1: Sampling Distributions
Unit 1: Basic Concepts of Sampling Distribution
Unit 7: Consistency
6
BLOCK 1: SAMPLING DISTRIBUTIONS
To draw inferences about the population characteristics (known as parameters) on the basis
of a sample, we require the sampling distribution of the statistic (function of sample
observations). This block of the course provides a brief discussion on the sampling distributions
of various statistics. It comprises five units. The detail on each unit is as follows:
Unit 1: Basic Concepts of Sampling Distribution make you familiar with the basic terms
required for understanding the sampling distribution. In this unit, you will study some important
terms such as population and sample, parameter and statistic, simple random sampling,
estimator and estimate, etc. The concept and role of sampling distributions in statistical
inference are described in this unit. The most important theorem “central limit theorem” and
law “law of large numbers” are also described with their applications.
One of the most important sample statistics which is used to draw a conclusion about the
population mean is the sample mean. In Unit 2: Sampling Distributions of Sample Means,
we present the sampling distributions of means (single mean and difference of two means)
under various situations. We describe the sampling distribution of mean when the population
is normally distributed or not. Also in this unit, the sampling distribution of difference of two
means when samples are independent and paired are discussed with examples.
Sometimes, we deal with the data collected in the form of counts and the collected data
classified into two categories or groups according to an attribute. In such situations, we use
sample proportion instead of mean. Similarly, in many practical situations, we have also been
concerned with variability. Especially when a little variance in diameter is crucial when
manufacturing things that fit together, like pipes and ball bearings. Therefore, in Unit 3:
Sampling Distributions of Sample Proportions and Variances, we discuss the sampling
distributions of sample proportion and difference of two proportions as well as the sampling
distributions of sample variance and ratio of two sample variances.
Unit 4: Sampling Distributions Associated with Normal Populations-I and Unit 5:
Sampling Distributions Associated with Normal Populations-II make you familiar with the
chi-square, t and F-distributions with their probability curve, summary measures, properties
and applications. In this unit, we also discuss how to read the tabulated values of these
sampling distributions.
Expected Learning Outcomes
After completing this block, you should be able to:
define statistical inference;
define the basic terms such as population and sample, parameter and statistic, simple
random sampling, estimator and estimate, etc. that are used in statistical inference;
explain the concepts of the sampling distribution and standard error;
describe the important theorem of sampling distributions “central limit theorem” and law
“law of large numbers” with their applications;
explain the sampling distributions of mean when the population is normally distributed
and not;
explore the sampling distributions of the difference of two means when samples are
independent and paired;
describe the shape of the sampling distributions of sample proportion and the difference
of two sample proportions; 7
explain the sampling distributions of sample variance and ratio of two sample variances;
explain the chi-square, t and F-distributions with their probability curve, summary
measures, properties, applications and relation to other distributions;
identify the conditions under which the chi-square, t and F-distributions can be used; and
describe the method of obtaining the tabulated value from the t, chi-square and F-
distribution tables.
a b : Beta function
B ( a,b ) =
a+b
a : Gamma function
t (n),(1− α) and t (n),α : The left and right-tailed tabulated values of the t-statistic with n
degrees of freedom
2
χ(n),(1− α) and χ(n),α : The left and right-tailed tabulated values of the chi-square statistic
2
8
UNIT 1
BASIC CONCEPTS OF
SAMPLING DISTRIBUTION
Structure
1.1 Introduction 1.6 Law of Large Numbers
Expected Learning Outcomes 1.7 Summary
1.2 Basic Terminology 1.8 Terminal Questions
1.3 Introduction to Sampling 1.9 Solutions /Answers
Distribution
1.4 Concept of Standard Error
1.5 Central Limit Theorem
1.1 INTRODUCTION
In many situations, we have to extract some information from all units or items
of a group/population. But, if
• the whole population is too large to study,
• the units of the population are destructive in nature,
• there are limited resources and manpower available, etc.
then gathering the information from all units is not practically inconvenient and
sometimes units are destroyed under investigation. For example, as you know
many of us use Facebook and if you are interested to know the average age of
the Facebook user, then you have to survey every person in the world who
uses Facebook. But it is not possible to survey everyone in the world.
Therefore, in most of the cases of daily life, business, industry, etc., the
information about the whole group/population is gathered through a random
sample. The results of a properly taken sample enable the investigator to
arrive at generalisations that are valid for the entire group/population. The
process of generalising sample results to the population is called Statistical
Inference.
To draw inferences about the population characteristics (known as
parameters) on the basis of a sample, we require the sampling distribution
of the statistic (function of sample observations). This unit as the title is
devoted to explaining the basic concepts required to understand the sampling
distribution as well as statistical inference.
9
Unit Writer- Dr. Prabhat Kumar Sangal, School of Sciences, IGNOU, New Delhi
Block 1 Sampling Distributions
This unit is divided into 9 sections. Section 1.1 is introductive. In Section 1.2,
we defined the basic terminology used in statistical inference such as
population and sample, parameter and statistic, simple random sampling,
estimator and estimate, etc. The concept and role of sampling distributions in
statistical inference are described in Section 1.3. In Section 1.4, you will study
the concept of standard error. The most important theorem “central limit
theorem” and law “law of large numbers” of Statistics are described with
their applications in Sections 1.5 and 1.6, respectively. The unit ends by
providing a summary of what we have discussed in this unit in Section 1.7 and
the terminal questions, and the solution of the SAQs/TQs are given in Sections
1.8 and 1.9, respectively.
In the next unit, we shall discuss the sampling distributions of sample means.
define the basic terms such as population and sample, parameter and
statistic, estimator and estimate, etc. used in statistical inference;
explain the concept of the law of large numbers with its application.
∑(X )
2
Here, we divide i − X by (n – 1) rather than n as our definition of the
i =1
variance. The reason for taking (n – 1) in place of n will become clear in Unit 6
of this course.
Statistical inference is a technique of drawing conclusions about the
population from data (sample). It is based on the assumption that the sample
should be random. We have various techniques to select a random sample. In
MST-013: Survey Sampling and Design of Experiments-I, you have studied
various sampling methods so that we can maintain the randomness. In
statistical inference, we use simple random sampling because for randomness
it is sufficient. Let us look at this sampling. 11
Block 1 Sampling Distributions
12
• Order matters. The order of selecting the typists in the sample is
Unit 1 Basic Concepts of Sampling Distribution
important. It means that if we select typist A first and then B (i.e. AB) differs
from selecting B first and then A (i.e. BA) and each of these samples is
regarded as a different possible sample that can be selected from this
group of typists.
• We replace the unit. A unit drawn at a time is replaced back to the
population before the subsequent draw. With replacement means a
sample of selecting the typist A and then A again (AA) is possible.
In this case, the total number of possible samples will be 16 which are listed
below:
For listing all possible
AA (repetition allows) CA(order matters) samples, we list samples
systematically. First, we list
AB CB(order matters)
all of the possible samples
AC CC (repetition allows) with the first element of the
AD CD population i.e. A as the first
BA (order matters) DA(order matters) typist, then all of the
possible samples with the
BB (repetition allows) DB(order matters)
second element of the
BC DC(order matters) population i.e. B, C and so
BD DD (repetition allows) on. In this way, we can be
sure that we have all of the
In this case, we can determine the total number of possible samples of any possible random samples.
size (n) that can be selected from a population of any size (N) using the
following formula:
Total number of possible samples = Nn .
The ordered sampling with replacement is also known as theoretical
sampling and it is used to develop the theories of sampling distribution. This
case is also known as simple random sampling with replacement
(SRSWR).
Unordered Sampling with Replacement
In this case,
• Order does not matter. The order of selecting the typists in the sample is
not important. It means that if we select typist A first and then B (i.e. AB) is
the same as selecting B first and then A (i.e. BA). These samples are
considered as one and are not considered separate samples.
• We replace the unit. A unit drawn at a time is replaced back to the
population before the subsequent draw. With replacement means a
sample of selecting the typist A and then A again (AA) is possible.
In this case, the total number of samples will be 10 which are listed as follows:
We can determine the total number of samples when order does not matter,
and repetition is allowed using the following formula:
Total number of possible samples = N+n −1
Cn = N+n −1
CN−1 .
AB CB(order matters)
AD CD
BC DC(order matters)
We can determine the total number of possible samples when order matters,
and repetition is not allowed using the following formula:
Total number of possible samples = N × (N − 1) × (N − 2) × ... × (N − n + 1) .
In this case, the total number of possible samples will be 6 which are listed as
follows:
14
Unit 1 Basic Concepts of Sampling Distribution
We can determine the total number of samples when order does not matter,
and repetition is not allowed using the following formula:
Total number of possible samples = N Cn .
We may select diverse samples. But in practice, order does not matter
because we do not care about the order in which units are selected. Also, we
usually do not allow one individual to be chosen twice. Therefore, we often do
the unordered sampling without replacement. Therefore, it is also called
experimental sampling. This is commonly called simple random sampling
without replacement (SRSWOR). However, in statistical inference ordered
simple random sampling with replacement (commonly called SRSWR) is used
to develop the theory of statistical inference. On the other hand, in a large
population, the probability of selecting the same individual twice is negligible,
and it can be demonstrated that the results obtained from sampling with
replacement are very close to the results obtained using sampling without
replacement. The main advantage of sampling with replacement is that the
sample observation will be independent, and this simplifies the analysis.
After understanding the various situations, let us take an example for
illustration purposes.
Example 1: There are five sales associates at a car showroom. The number
of cars they sold last week is as follows:
Vihaan 2
Rohan 5
Ritika 4
Hassan 6
Anita 10
Sample Sample
Sample Sample in Term Sample Sample in Term
Observations Observations
Number of Associates Number of Associates
(car sold) (car sold)
1 (Vihaan, Vihaan) (2, 2) 14 (Ritika, Hassan) (4, 6)
For listing all possible
2 (Vihaan, Rohan) (2, 5) 15 (Ritika, Anita) (4, 10)
samples, we list samples
systematically. First, we list 3 (Vihaan, Ritika) (2, 4) 16 (Hassan, Vihaan) (6, 2)
all of the possible samples
4 (Vihaan, Hassan) (2, 6) 17 (Hassan, Rohan) (6, 5)
with the first element of the
population i.e. A as the first 5 (Vihaan, Anita) (2, 10) 18 (Hassan, Ritika) (6, 4)
typist, then all of the
possible samples with the 6 (Rohan, Vihaan) (5, 2) 19 (Hassan, Hassan) (6, 6)
second element of the 7 (Rohan, Rohan) (5, 5) 20 (Hassan, Anita) (6, 10)
population i.e. B, C and so
on. In this way, we can be 8 (Rohan, Ritika) (5, 4) 21 (Anita, Vihaan) (10, 2)
sure that we have all of the 9 (Rohan, Hassan) (5, 6) 22 (Anita, Rohan) (10, 5)
possible random samples.
10 (Rohan, Anita) (5, 10) 23 (Anita, Ritika) (10, 4)
SAQ 1
A hospital administrator wishes to estimate the mean weight of babies born in
her hospital. She collects the birth records of a day and the observes weights
(in pounds) of 4 babies (B1, B2, B3, B4) as 6, 8, 7, and 6 pounds. How many
samples of size 2 are possible with replacement? Also, write them.
After understanding the concept of population, sample and how many samples
are possible in different cases, let us move to the next concept.
Parameter
The characteristics of a population can be described with some measures
such as population mean, variance, etc. These measures are known as
parameters of the population. Thus, we can define a parameter as:
“A parameter or population parameter is a numerical value that
summarises or measures or represents a specific characteristic of an
entire population.”
The parameters are derived from data collected from the entire population and
taken as fixed constants.
For example, suppose the course coordinator of the MST-016 course
calculates the average marks of all the learners in the MST-016 course then
the obtained average mark is a parameter because it is based on all learners
16
of the MST-016 course. Similarly, population variance, population coefficient of
Unit 1 Basic Concepts of Sampling Distribution
variation, population correlation coefficient, etc. are all parameters. Population
parameters are usually denoted by Greek letters, such as the population mean
and variance are represented by Greek letters µ and σ2, respectively.
Generally, the parameters of a population are typically unknown and are
estimated from sample data.
We know that the population of measurements such as height, marks, etc can
be described with the help of distribution such as normal, Poisson, etc. and the
distribution is fully determined with the help of its constants such as, in case of
a normal distribution, we need to know µ and σ2 to determine the normal
distribution, in case of Poisson distribution, we need to know λ, etc. These
constants are also known as the parameters.
Statistic
As a parameter describes the characteristic of the population in a similar way
a statistic describes the characteristic of the sample. We can define a statistic
as:
“A sample statistic or statistic is a numerical measure that summarizes
or describes a characteristic of a sample.”
A statistic is calculated using sample data. For example, suppose the course
coordinator of the MST-016 course calculates the average marks of the
learners in the MST-016 course by selecting some learners randomly instead
of all learners then the obtained sample average mark is a statistic because it
is based on the sample of the learners of the course. Similarly, sample
variance, sample coefficient of variation, sample correlation coefficient, etc.
are all statistics (plural of statistic). The statistic is usually denoted by
Roman/Latin letters, such as the sample mean and variance are represented
by X and S2, respectively. Generally, the statistic is used to estimate the
population parameter. We may also define a statistic as
Any quantity calculated from sample values that does not contain any
unknown parameter is known as a statistic.
For example, if X1 , X2 , ..., Xn is a random sample of size n taken from a
population with mean µ and variance σ2 (both are unknown) then the sample
1 n X−µ
mean X = ∑ Xi is a statistic whereas and X / σ are not statistics
n i=1 σ/ n
because both are functions of unknown parameters. If both µ and σ2 are
X −μ
known then or X / σ become a statistic because both µ and σ2 are
σ/ n
known and are treated as constants. We use different symbols for parameters
and statistics as follows:
Variance σ2 S2
Proportion P p
17
Block 1 Sampling Distributions
After understanding the concept of basic terminology, you are now ready to
learn the main concept of this unit, that is, sampling distribution. Let us discuss
it in the next section.
18
Unit 1 Basic Concepts of Sampling Distribution
From the table, the average (mean) cholesterol level of these selected
persons is 195 mg/dl. Now, if you use this sample to make an inference about
the population's average cholesterol level then we say that the sample
average cholesterol level of 195 mg/dl is an estimate of the average
cholesterol level of the persons living in this city (population). However, we do
not know the precision (the estimate is accurate or not) of the estimate
because the actual cholesterol level of the persons in the city is unknown.
If another random sample of 10 persons (Sample-II) is selected and measured
their cholesterol levels (shown in column 3 of Table 1.2) then we get the
average (mean) cholesterol level of these selected persons is 202 mg/dl. This
is different from Sample-I because it contains different persons who have
different cholesterol levels. Thus, if you use this sample mean to estimate the
population mean then you get a different estimate of the average cholesterol
level for the whole population. If other samples of 10 persons are selected, it is
unlikely that exactly the same mean would be observed, therefore, we may
expect different estimates of the population mean every time. Now, the
following questions may arise:
• How well does a sample describe its population?
• How can we tell which sample gives the best estimate of the population
parameter?
• What is the probability of selecting a sample with specific
characteristics?
These questions can be answered once we establish the sampling
distribution of the statistic such as mean, variance, proportion, etc. The
sampling distribution of a statistic is important because it enables us to draw
conclusions about the corresponding population parameter.
For a better understanding of the concept of the sampling distribution, we
consider a population of very small size so that we can easily obtain unknown
population parameters, say, mean (µ) and standard deviation (σ) using all
population observations. And see how the sampling distribution helps us to
draw conclusions about the population.
Consider a small production industry which has five employees. The monthly
salary (in thousands) of each employee is given as follows:
Table 1.3: Monthly Salaries of Employees
Monthly Salary
Employee
(in thousands)
Lavnik 25
Avishi 30
Aman 15
Tanishq 25
Harsh 10
1 N (25 − 21)2 + (30 − 21)2 + (15 − 21)2 + (25 − 21)2 + (10 − 21)2
σ
= ∑ (Xi − =
N i=1
μ)2
5
16 + 81 + 36 + 16 + 121
= = 54 7.35
=
5
We can plot the graph of the population (monthly salary) to know the form of
the population as follows in Fig. 1.1.
From the above figure, we observe that the monthly salary of the employees
does not follow a known distribution (especially normal) and it is left-skewed.
Let us assume that we do not know the average salary of the employees. So
we decide to estimate the population mean on the basis of a sample of size
n = 2. As you have studied in Section 1.2, there are Nn = 52 = 25 possible
simple random samples with replacement of size 2. All possible samples of
size n = 2 are given in the second column of Table 1.4. We also calculate the
means of each sample which are given in the last column of the same table.
Table 1.4: Samples and Sample Means
From the above table, you can see that the value of the sample statistic
(sample mean) varies from sample to sample and out of 25 samples there is
no sample whose mean is equal to the population mean. It means that out of
25 samples, no one estimates the population mean exactly. So, what do we do
now? How can we estimate it? The sampling distribution helps us in such
situations. We now try to understand the concept of it.
Since the sample mean varies from sample to sample, therefore, it is treated
as a random variable. As you have studied in MST-012, a random variable has
a distribution, so the sample statistic (sample mean) has a distribution. To
obtain the distribution of a statistic (sample mean) we arrange the values of it
in ascending or descending order and calculate the frequency of each value
as shown in Table 1.5. We can also obtain the probability distribution (using
the relative frequency approach of probability) described in MST-012 of the
occurrence of each value which is given in the last column of Table 1.5.
Table 1.5: Sampling Distribution of Sample Means
So the arrangement of all possible values of the sample mean with their
corresponding probabilities is called the sampling distribution of mean. Thus,
22 we can define the sampling distribution in general as follows:
Unit 1 Basic Concepts of Sampling Distribution
“The probability distribution of all possible values of a sample statistic
that would be obtained by drawing all possible samples of the same size
from the population is called the sampling distribution of that statistic.”
To get an idea of the shape of the sampling distribution of mean, we plot the
graph (frequency bar) of the values of the sample mean taking the sample
mean on the X-axis and corresponding frequencies on the Y-axis as shown in
Fig. 1.2(b).
Distribution
of all the
possible
sample
means (n = 2)
from this
population.
From Fig. 1.2, we observe that the shape of the sampling distribution has been
changed.
As the probability distribution such as normal, Poisson, binomial, etc. allows us
to gain an understanding of the summary (mean, standard deviation, etc),
likelihood and probabilities of different values occurring in the outcome, in a
similar way, a sampling distribution is a probability distribution that does the
same job. The sampling distribution of the sample means itself has mean,
variance, etc. Therefore, we can calculate the mean of sample means as
1 k k
Mean of the sample means = (X) = ∑ Xi fi where, K = ∑ fi
K i=1 i =1
1
= (10 × 1 + 12.5 × 2 + ... + 30 × 1) = 21= μ
25
We can also calculate the mean of the sample means as
k
1 2 1
E(X) =X =∑ Xipi =10 × + 12.5 × + ... + 30 × =21
i =1 25 25 25
You saw! The mean of the sampling distribution of means is equal to the
population mean. This is what we wanted. However, we did not get the same
using a single sample. We find it as the mean of the sampling distribution of
means.
Thus, we can say that to find the exact estimate of the unknown population
parameter, we first find the sampling distribution of the corresponding statistic
and then compute the mean of the obtained sampling distribution of that
statistic.
We now calculate the standard deviation of the sampling distribution of mean
as
23
Block 1 Sampling Distributions
1 k
( )
2
SD(X)
= ∑ fi Xi − X
K i=1
1
( )
SD X =
25
1× (10 − 21) + 2 × (12.5 − 21) + ... + 1× ( 30 − 21)
2 2 2
1
= (121 + 144.5 + ... +=
121) 5.20
25
( )
SD X = 5.20
Sample
Sample Sample in Term of
Observation Sample Mean
number Employees
(monthly salary)
10 1 1/125 = 0.008
15 7 7/125 = 0.056
20 15 15/125 = 0.12
25 11 11/125 = 0.088
30 1 1/125 = 0.008
1
= (10 × 1 + 11.67 × 3 + ... + 30 × 1) = 21= μ
125
1 k
( )
2
SD(X)
= ∑ fi Xi − X
K i=1
1
SD X=( ) 125
1× (10 − 21) + 3 × (11.66 − 21) + ... + 1× ( 30 − 21) =
2 2 2
4.24
We now plot the sampling distribution of mean when the sample size n = 3 in
Fig. 1.3.
From Fig. 1.3, you can observe that it is similar to the normal distribution.
From the above discussion, we observe some features of the sampling
distribution:
The shape of the sampling distribution of mean may be quite different
from the shape of the population.
The sample means “pile up” around the population mean and “tail off”
towards the extremes. For this example, the population mean is µ = 21,
and the sample means are clustered around a value of 21. It should not
surprise you that the sample mean tends to approximate the population
mean.
The sampling distribution of sample mean tends to bell-shaped normal
probability distribution as sample size n increases. 25
Block 1 Sampling Distributions
SAQ 3
A Municipal Corporation Office has four typists. An officer gave the same
sample page of a manuscript to all four typists. The number of errors made by
each typist is shown in Table 1.7 as given below:
Table 1.7: Number of Errors per Typist
A 4
B 2
C 3
D 1
• How can know the precision (the estimate is accurate or not) of the
estimate when it is not possible to compute the exact value?
The standard error does such jobs for us. It measures how far the sample Standard deviation and
mean (average) of the data is likely to be from the true population mean. The standard error of the mean
standard error provides a measure of how much distance is expected on are both statistical
average between a sample mean and the population mean. It gives the measures of variability.
accuracy of a sample mean by measuring the sample-to-sample variability of While the standard
the sample means. We now formally define the standard error (SE) as deviation of a sample
depicts the spread of
“The standard deviation of a sampling distribution of a statistic (mean,
observations within the
proportion, standard deviation) is known as standard error”.
given sample regardless of
The standard error is not denoted by any symbol whereas it is denoted by its the population mean where
abbreviation SE. as the standard error of the
mean measures the degree
The standard error is a standard deviation of the sampling distribution, so it
of dispersion of sample
serves the same two purposes for the sampling distribution as the standard
means around the
deviation of the data which are listed as follows:
population mean.
1. The standard error provides a measure of how much discrepancy is
expected from one sample to another. When the standard error is small,
then all of the sample means are close together and have similar values. If
the standard error is large, then the sample means are scattered over a
wide range and there are major differences from one sample to another.
1 k
( )
2 k
SE(X)
= ∑ fi Xi − X where, K = ∑ fi
K i=1 i =1
1
( )
SE X =
25
1× (10 − 21) + 2 × (12.5 − 21) + ... + 1× ( 30 − 21)
2 2 2
1
= (121 + 144.5 + ... +=
121) 5.20
25
1 n
( )
2
=S ∑ Xi − X
n − 1 i=1
These are the expressions of the standard error of the mean. We give the
expressions of standard error for other statistics such as proportion, standard
deviation, etc. in the subsequent units.
28
Unit 1 Basic Concepts of Sampling Distribution
Let us try one example relating to standard error.
Example 2: The diameter of a steel ball bearing produced by a semi-
automatic machine is known to be distributed normally with a mean of 12 cm
and a standard deviation of 0.1 cm. If we take a random sample of size 10
then find the standard error of the sample mean for estimating the population
mean of the diameter of all ball bearings produced.
Solution: Here, we are given that
μ = 12, σ = 0.1, n = 10
Since, population standard deviation (σ) is given, therefore, we can calculate
the standard error of the sample mean as
σ 0.1
SE =
X ( ) =
n
= 0.03
10
From the above formula, the magnitude of the standard error depends on two
factors:
1. The standard deviation of the population from which the sample is
selected; and
2. The size of the sample.
We examine each of these factors one at a time. 29
Block 1 Sampling Distributions
9 4 1.33
9 9 3
9 16 5.33
9 25 8.33
9 36 12
9 49 16.33
9 64 21.33
9 81 27
9 100 33.33
We now plot the standard error corresponding to the standard deviation of the
population as shown in Fig. 1.4.
Fig. 1.4: The standard error with the population standard deviation.
From the above figure, we observe that as the standard deviation of the
population decreases, the standard error or the average distance of the
sample mean deviating from the population mean also decreases.
Effect of sample size (n)
As we noted in the previous section, as we increase the sample size n from 2
to 3 then standard error decreases. Also, from the formula for calculating
standard error, we observe that there is an inverse relationship between the
sample size and the standard error. As we increase the sample size, the
30
Unit 1 Basic Concepts of Sampling Distribution
standard error decreases. It means that if we collect more data in a sample,
our estimate of the population mean will be more accurate. To illustrate the
general relationship between standard error and sample size, we visualize the
effect of the sample size in Fig. 1.5. For that, we select samples of different
sizes from a single population with a standard deviation equal to 9 and
compute the standard error for each sample as follows:
1 9 9
4 9 4.5
9 9 3
16 9 2.25
25 9 1.8
36 9 1.5
49 9 1.3
64 9 1.13
81 9 1
100 9 0.9
We now plot the standard error corresponding to the sample size as shown in
Fig. 1.5.
From the above figure, we observe that the standard error of the mean will
approach zero with the increasing number of observations in the sample. It
happens because when the sample size becomes large then the sample
becomes more and more representative of the population, and the sample
mean approaches the actual population mean.
Note: The standard errors discussed above were calculated under the
assumption that sampling is done either from an infinite population or from a
finite population with replacement. But, in real-life sampling problems, most
sampling plans do not permit an element to be selected twice in a given
sample (i.e. sampling with replacement). Consequently, if the population is not
large in comparison to the size of the sample and sampling is done without
replacement then we multiply the above standard errors by the correction
N−n
factor .
N −1 31
Block 1 Sampling Distributions
In practice, if the sample size n is less than or equal to 10% of the population
size N, this correction factor may be ignored for all practical purposes.
You should try the following Self Assessment Question before studying next
section.
SAQ 4
A juice company packages the juice in 500 mL pouches. Suppose you are
incharge of monitoring that pouches that they are being filled correctly and you
randomly select a sample of 25 pouches from the thousands of pouches filled
during a shift. Given that the standard deviation of the juice-packaging process
is 12 mL,
(ii) If you select a sample of 100 pouches, what will be the standard error?
(iii) What will be the size of the standard error of the mean as the sample
size is increased from 25 to 100?
We now introduce you to some of the most powerful data laws, principles, and
rules that are based on statistical theories. These will help you become a
better data scientist or data analyst in general. The central limit theorem and
law of large numbers are the foundation of statistical inference. Understanding
these fundamental concepts can help you analyse and interpret data more
effectively, enabling you to make confident decisions based on solid evidence.
In the coming session, we will discuss the central limit theorem and in the next
session, the law of large numbers.
The first concept of the central limit theorem was given by French
(1667-1754) Mathematician Abraham De Moivre in 1733. At that time, the central limit
Abraham De Moivre was a theory proposed by De Moivre was not popular. But another well-known
French mathematician. He French mathematician, Pierre-Simon Laplace, revived the idea in 1812. The
was the first who gave the central limit theorem discoveries made by Laplace at that time caught the
central limit theorem. interest of numerous academics and thinkers. The central limit theorem was
extended later in 1901 by Russian mathematician Aleksandr Lyapunov. He
32
Unit 1 Basic Concepts of Sampling Distribution
disproved the idea in general and provided mathematical evidence for its
validity.
The central limit theorem is one of the most important theorems of Statistics. It
states that
The sampling distribution of the mean approaches to a normal
distribution as the size of the sample increases, regardless of the shape
of the original population distribution.
We can explain the central limit theorem as
If a random sample of size n is taken from a population with mean µ and finite We can apply the central
variance σ2 then the sampling distribution of the sample mean tends to a limit theorem to almost all
normal distribution with mean µ and variance σ2/n as the sample size tends to types of probability
be large (n ≥ 30) whatever may be the form of the parent population, that is, distributions, but the
population must have a
σ2
X ~ N μ, when n ≥ 30 finite variance. That
n restriction rules out the
Cauchy distribution
We do not intend to prove this theorem here but merely show graphical
because it has infinite
evidence of its validity in Fig. 1.6. Here, we will also try to show how large
variance.
must the sample size be for which we can apply the central limit theorem.
From the first column of this figure, we observed that when the parent
population is normal then all sampling distributions of mean for varying sample
sizes n = 2, 5 and 30 are also normal, having the same mean but their
variances decrease as n increases.
The second column of this figure represents the uniform population. Here, we
observe that the sampling distribution of mean is symmetrical and does not
follow any standard distribution when n = 2 and tends to be normal when
n = 30.
However, the third column of this figure represents the binomial population
(discrete). Again, when n = 2, the sampling distribution of mean is symmetrical
and for n = 5 it is quite bell-shaped and tends to be normal when n = 30.
The last column of this figure represents the exponential population which is
highly skewed. Here, we also observe that when the sample size n = 2 then
the sampling distribution of mean does not follow any standard distribution but
for n = 30, the distribution of mean is symmetrical bell-shaped normal.
From Fig. 1.6, we also observe that the rate at which the distribution
approaches a normal distribution depends on the shape of the population. If
the population itself is normally distributed, the sampling distribution of mean is
also normal for any sample size n, as stated earlier. On the other hand, for
33
Block 1 Sampling Distributions
Fig. 1.6: Sampling distribution of sample means for various populations when n = 2,
n = 5 and n = 30.
34
Unit 1 Basic Concepts of Sampling Distribution
Here, we also conclude that if we draw a random sample of large size n ≥ 30
from the population then the sampling distribution of mean can be
approximated by a normal probability distribution, whatever the form of the
parent population.
From the above discussion, we draw the following results:
1. When the population/study variable is normally distributed, then the
sampling distribution of mean will be normally distributed, for any sample
size n.
2. When the distribution of the population/study variable is not normal, a
sample size of 30 or more is needed to use a normal distribution to
approximate the distribution of the sample means.
3. If the population distribution is fairly symmetrical, the sampling distribution
of the mean is approximately normal for samples as small as 5.
Let us see the applications of the central limit theorem using an example.
Example 3: The average breaking strength of a certain brand of steel cable is
2500 pounds, with a standard deviation of 160 pounds. A sample of 40 cables
is randomly selected and tested. What is the sampling distribution of mean?
Also, find the probability that the average breaking strength of the sample
cables:
(i) more than 2550 pounds
(ii) less than 2480 pounds
(iii) between 2450 pounds and 2550 pounds
( )
E X = μ= 2500
and variance
σ 2 (160)2
( )
Var X
= =
n 40
= 640
(i) The probability that the average breaking strength of the sample cables
is more than 2550 pounds is given by
To get the value of this probability, we convert the sample mean X into
a standard form. Since the sampling distribution of the average breaking
strength (X) of the steel cables is normally distributed with mean of 2500
pounds and a variance of 640 pounds2, therefore, we transform
the sample mean (X) into the standard normal Z-score as
Therefore, subtract 2500 from each term and then divide each term by
25.30 in the above inequality. Thus, the probability expression becomes
X − 2500 2550 − 2500
P > =P [ Z > 1.98]
25.30 25.30
1 − P [ Z ≤ 1.98]
=
X 2500
= X 2550 From the standard normal table (Table IV) given in the Appendix of this
Z 0=Z 1.98 volume, we get
Fig. 1.7 1 − P [ Z ≤ 1.98] = 1 − 0.9761 = 0.0239
P X > 2550 =
(iii) We can compute the probability that the average breaking strength of the
sample cables lies between 2450 pounds and 2550 pounds as
SAQ 5
A survey of a city found that the mean number of days per month that people
who suffer from migraine headaches is 12.4 days with a standard deviation 3
days. Find the probability that if a random sample of 50 people who suffer from
migraine headaches is selected, the mean of the sample will be between 12
and 13 days.
This is the same intuition behind the idea that if we collect more data, our
sample of data will be more representative of the population.
37
Block 1 Sampling Distributions
Here, try to understand the law of large numbers in action with the help of an
example.
Suppose the distribution of the weight of all the young men living in a city is
close to a normal distribution with a mean weight of 65 kg and a standard
deviation of 5 kg. To understand the law of large numbers, we calculate the
sample mean weight ( X ) for varying sample sizes n = 1, 2, 3, …. Fig. 1.10
shows the behaviour of the sample mean weight X of men chosen at random
from this city. The graph plots the values of X along the vertical axis and
sample size along the horizontal axis as sample size varying n =1, 2, 3…
First, we start with a sample of size n = 1, that is, we select a man randomly
from the city. Suppose the selected man has a weight of 70 kg, therefore, the
line of the graph starts from this point. We now select the second man
randomly and suppose his weight is 62 kg. So for n = 2, the sample mean is
70 + 62
=X = 66 kg
2
This is the second point on the graph. Now, we select the third man randomly
from the city and suppose his weight is 55 kg. Therefore, for n = 3, the sample
mean is
70 + 62 + 55
=X = 62.33 kg
3
This is the third point on the graph.
This process will continue. From the graph, we can see that the mean of the
sample changes as we make more observations, eventually, however, the
mean of the observations gets closer and closer to the population mean of 65
kg as the sample size increases.
Hence from Fig. 1.10, we can observe that every additional data point
gathered has the potential to move the sample to the true population mean.
We use the law of large numbers in a variety of fields including statistics,
market research, finance, healthcare, insurance, engineering, etc. Here we
mention some examples of the use of the law of large numbers:
1. In market research, let us suppose you are working for a marketing
38 company, and you want to know what percentage of people in a certain city
Unit 1 Basic Concepts of Sampling Distribution
prefer your brand of mobile. It is not possible for you to survey every single
person in the city, so you take a random sample of 100 people and ask
them which brand of mobile they prefer. If you find that 40% of them prefer
your brand. Does that mean that 40% of the entire population prefers your
brand? Not necessarily there is always a chance that your sample was
biased in some way.
However, if you were to take a larger sample, say, 10,000 people and found
that 45% of them prefer your brand, that would be a stronger indication that
a majority of the population really does prefer your brand. And if you were
to take an even larger sample say, 100,000 people and found that 48% of
them prefer your brand, that would be an even stronger indication that your
brand is really popular in the city.
2. In financial analysis, we can also use the law of large numbers. For
example, suppose you are a stock analyst and you want to know the
average rate of return for a certain stock. You could calculate the rate of
return for a small sample of investors and get a rough estimate, but it would
not be very accurate. However, if you take a larger sample of investors to
calculate the rate of return, you will get a more accurate estimate of the true
rate of return.
3. Similarly, the law of large numbers is also used in medicine to study the
effectiveness of treatments. If you want to know whether a certain
medication is effective for a particular condition then you could study a
small sample of patients and get a rough idea. However, if you study a
larger sample of patients, you will get a more accurate estimate of the
effectiveness of the medication.
After studies where we can use the law of large numbers, we should note the
following important points:
• The law of large numbers does not tell that the sample mean will always
reflect the true population characteristics, especially for small samples.
• If a given sample mean deviates from the true population mean, then the
law of large numbers does not guarantee that successive samples will
move the observed average towards the population mean.
I hope you have understood the concept of the law of large numbers and how
you can apply it to real-world situations. Hence, you can make better decisions
and draw more accurate conclusions from your data.
With the help of the law of large numbers, we can also determine the minimum
sample size to get a reliable inference about the population.
This law states that for any two arbitrary small numbers ɛ (ε > 0) and ɳ (0 < η
σ2 You have also studied the
< 1) , there exists an integer n = 2 such that if a random sample of size n or
εη law of large numbers in
Unit 18 of MST-012:
larger is drawn from the population and the sample mean (X) is calculated for
Probability and Probability
this sample, then the sample mean (X) arbitrarily close to the population
Distribution. Where you
mean (µ) where σ2 is the finite variance of the population. In statistical studied the same in the
inference ɛ is called margin of error and 1 – η is called confidence level. context of probability.
Hence, to get a reliable inference about the population, we can determine the
39
Block 1 Sampling Distributions
\minimum sample size with the help of the law of large numbers. For that, the
σ2
minimum sample size should be n ≥ 2 .
εη
We can calculate the minimum sample size for estimating the mean weight of
babies born in her hospital as
( 0.5 )
2
σ2
n≥ 2 = 2
ε η ( 0.4 ) × 0.01
= 156.25 ~ 157
Hence, the hospital administrator should take a random sample of at least 157
babies.
Now, try the following Self Assessment Question for your practice.
SAQ 6
A pathologist wants to estimate the mean time required to complete a certain
analysis on the basis of a sample study so that he may be 99% confident that
the mean time may remain within ± 2 days of the mean. As per the available
records, the population standard deviation is 5 days. What must be the size of
the sample for this study?
We now end this unit by giving a summary of what we have covered in it.
1.7 SUMMARY
In this unit, we have covered the following points:
• The statistical procedure which is used for drawing conclusions about the
population parameter on the basis of the sample data is called “statistical
inference”.
• A population is a group of measurements in the quantitative or qualitative
form of the characteristic under study.
• A parameter or population parameter is a numerical value that summarises
or measures or represents a specific characteristic of an entire population.
• A sample statistic or a statistic is a numerical measure that summarizes or
describes a characteristic of a sample.
40
• Any statistic used to estimate an unknown population parameter is known
Unit 1 Basic Concepts of Sampling Distribution
as “estimator” and the particular value of the estimator is known as
“estimate” of the parameter.
• The probability distribution of all possible values of a sample statistic that
would be obtained by drawing all possible samples of the same size from
the population is called the “sampling distribution” of that statistic.
• The standard deviation of the sampling distribution of a statistic is known
as “standard error”.
• The central limit theorem states that the sampling distribution of the
sample means tends to a normal distribution as the sample size tends to
be large (n > 30).
• The law of large numbers states that as a sample size increases, the
sample mean gets closer to the average of the whole population.
1 n
( )
2
=S ∑ Xi − X
n − 1 i=1
(29 − 30)2 + (31 − 30)2 + (30 − 30)2 + (32 − 30)2 + (30 − 30)2
+(29 − 30)2 + (30 − 30)2 + (30 − 30)2 + (29 − 30)2 + (30 − 30)2
=
9
1+ 1+ 0 + 4 + 0 + 1+ 0 + 0 + 1+ 0 8
= = = 0.94
9 9
In this case, we are estimating the population standard deviation from
the sample standard deviation, therefore, the sample standard deviation
is the estimator for the population standard deviation and 0.94 pounds is
the estimated value of the population standard deviation.
3. First of all, we check the shape of the population. For that, we pot the
population values (number of errors) as follows:
Sample Sample in
Number Sample Sample Mean
Terms of
Observation (X)
Typist
X Frequency(f) Probability(p)
1
= (1.0 × 1 + 1.5 × 2 + ... + 4.0 × 1=) 2.5= μ
16
Similarly, we can calculate the standard deviation of the sample
distribution of mean as
1 k k
∑ fi ( xi − μ) where, K = ∑ fi
2
SD(X)
=
K i=1 i =1
1
1× (1.0 − 2.5 ) + 2 × (1.5 − 2.5 ) + ... + 1× ( 4.0 − 2.5 )
2 2 2
=
16
1 10
= ( 2.25 + 2 + ... + 2.25=) = 0.791
16 16
Hence, we can conclude that the mean of the sampling distribution of
mean is the same as the population mean whereas the dispersion of the
sampling distribution (0.791) is less in comparison to the population
(1.12).
( )
E X = μ= 12.4 days
and variance
σ2
( )
Var X =
n
(3)2
= = 0.18
50
We have to find the probability that the sample mean will lie between 12
and 13 days is given by
P 12 < X < 13 [see Fig. 1.13]
Therefore, subtract 12.4 from each term and then divide each term by= X 2450
= X 2500
= X 2550
0.42 in the above inequality. Thus, the probability expression becomes: Z=−1.98 Z =
0 Z=1.98
Fig. 1.13
12 − 12.4 X − 12.4 13 − 12.4
P < < = P [ −0.95 < Z < 1.43]
0.42 0.42 0.42
It means that 75.25% of the mean of the sample will lie between 12 and 45
Block 1 Sampling Distributions
13 days.
6. Here, we are given
1−= η 0.01, ε = 2 , σ = 5
η 0.99 ⇒ =
46
UNIT 2
SAMPLING DISTRIBUTIONS OF
SAMPLE MEANS
Structure
2.1 Introduction When Populations are Normally
Distributed
Expected Learning Outcomes
When Populations are not Normally
2.2 Sampling Distribution of
Distributed
Mean
2.7 Sampling Distribution of
2.3 Sampling Distribution of
Difference of Two Means
Mean When Population is
When Samples are Paired
Normally Distributed
When Population of Differences is
When Population Variance is
Normally Distributed
Known
When Population of Differences is
When Population variance is
not Normally Distributed
Unknown
2.8 Summary
2.4 Sampling Distribution of
Mean When Population is 2.9 Terminal Questions
Not Normally Distributed
2.10 Solutions/Answers
2.5 Sampling Distribution of
Difference of Two Means
2.6 Sampling Distribution of Tools You Will Need
Unit Writer- Dr. Prabhat Kumar Sangal, School of Sciences, IGNOU, New Delhi
Block 1 Sampling Distributions
Child Name Avaya Ishaan Ria Zara Kavya Parth Rayaan Nyra Yash
Score 2 5 4 3 1 3 2 4 3
Here, you can easily calculate the mean and variance of the test scores
(population) as we have calculated in Unit 1. We get mean = 3 and variance =
1.33. To know whether the shape of the population (test scores) is normal or
not, we first plot the graph of the scores of the children as shown in Fig. 2.1.
Fig. 2.1: The distribution of the scores of the children in the listening task.
From the above figure, you can observe that the test scores of the children
follow an approximate bell-shaped normal distribution.
To obtain the shape of the sampling distribution of mean in this case, let us
consider all possible samples with replacement of size n = 2 from the above
population of the test scores of the children. There are Nn = 92 = 81 possible
simple random samples with replacement of size 2 which are given in the
second column of Table 2.1. We also calculate the mean of each sample
50 which are given in the last column of the same table.
Unit 2 Sampling Distributions of Sample Means
Table 2.1: Samples and Sample Means
Here, we have given some of these possible samples to reduce the space.
You can prepare the same as we have discussed in Unit 1. To obtain the
distribution of the sample means, we arrange the values of the sample mean
in ascending order and calculate the frequency corresponding to each value
as discussed in Unit 1. The obtained sampling distribution is shown in Table
2.2.
Table 2.2: Sampling Distribution of Means
1 1 1 1/81 = 0.012
3 2 10 10/81 = 0.123
5 3 19 19/81 = 0.235
7 4 10 10/81 = 0.123
9 5 1 1/81 = 0.012
Total 81 1
To get an idea of the shape of the sampling distribution of mean, we plot the
graph (frequency bar) of the values of the sample mean taking the sample
mean on the X-axis and corresponding frequencies on the Y-axis as shown in
Fig. 2.2(a).
From Fig. 2.2(a), you can observe the shape of the sampling distribution. You
may notice that even for a small sample size n = 2, the sampling distribution of
sample means when samples are taken from the normally distributed
population and population variance is known is normally distributed. 51
Block 1 Sampling Distributions
Let us see what happens as we increase the sample size. We now select all
possible samples with replacement of size n = 3 instead of 2 and then
calculate the sample mean for each sample and also prepare the graph to
observe the shape of the sampling distribution as discussed. The graph of the
values of the sample means is shown in Fig. 2.2(b).
Fig. 2.2: Sampling distribution of mean when population is normal for sample size n = 2,
3, 5, 30.
From Fig. 2.2(b), you can see that the sampling distribution of mean is also
normal. Similarly, we can prepare the sampling distribution of various sample
sizes and plot the graphs. We also prepared the graphs for n = 5 and 30 in
Fig. 2.2 (c) and (d), respectively. From Fig 2.2 (c) and 2.2(d), you can see that
the sampling distribution of mean is also normal, and the sample means are
distributed more tightly around the population mean as the sample size is
increased.
After knowing the shape of the sampling distribution of mean when the
population is normally distributed and population variance is known, you may
be interested to know the mean and variance of the sampling distribution.
Let us find the mean, variance and standard error of the sampling distribution
of mean.
In practice, only one random sample is selected, and the concept of the
sampling distribution is used to draw the inference about the population
52 parameters. If X1, X2, …, Xn is a random sample of size n taken from a normal
Unit 2 Sampling Distributions of Sample Means
population with mean µ and known variance σ then we can obtain the mean
2
X + X2 + ... + Xn
Mean=
of X E=
X E 1
( )
n [by definition of mean]
1
= [E(X1 ) + E(X2 ) + ... + E(Xn )] E[X ± Y]= E [ X] ± [ Y ]
n
Since X1, X2, …, Xn are randomly drawn from the same population, so they
also follow the same distribution as the population. Therefore,
E(X1=
) E(X2=
) ...
= E(Xn=
) μ
and
Var(X1=
) Var(X2=
) ... ) σ2
= Var(Xn=
Thus,
1 1
E(X)= n ( nμ) μ
μ + μ + ... + μ= =
n
n − times
( )
E X =μ
and variance
If X and Y are independent
1 1
2 [
= Var (X1 + X2 + ...=
Var(X)
+ Xn ) Var(X1 ) + Var(X2 ) + ... + Var(Xn )] random variables and a and
n n b are two constants then
2 Var ( aX + bY )
1 2 1
σ + σ2 = n2 ( nσ )
+ ... + σ 2
=
n2 = a2 Var ( X ) + b2 Var ( Y )
n − times
σ2
( )
Var X =
n
Hence, we conclude that if the samples are drawn from a normal population
with mean µ and known variance σ2 then the sampling distribution of mean X
is also normally distributed with mean µ and variance σ2/n, that is,
σ2
( )
If Xi ~ N μ,σ 2 then X ~ N μ,
n
and the quantity
X −μ
Z= ~ N ( 0,1) The standard deviation of a
σ
sampling distribution of a
n statistic (mean, proportion,
follows the standard normal distribution. standard deviation) is
known as standard error.
In other words, we can say that if we take all possible samples of a constant
size from a normal population and compute a statistic Z for each sample, then The standard error of the
the sampling distribution of Z would be standard normal. mean measures the degree
of dispersion of sample
The standard error of sample mean can be obtained by the definition of means around the
standard error as population mean.
σ
SE
= X =( )
SD X ( ) Var
= ( )
X
n 53
Block 1 Sampling Distributions
The above discussion was based on the assumption that the population was
infinite or extremely large or we were drawing samples with replacement. But if
the population is finite and we use sampling without replacement then a
statistical adjustment can be made. The adjustment is called the finite
correction factor (FPC). Without it, the central limit theorem does not hold
under those sampling conditions, and the standard error of the mean will be
too big. Under these conditions, the sample mean is distributed normally with
mean
( )
E X =μ
and variance
N − n σ2
( )
Var X =
N −1 n
The proof of this formula is beyond the scope of this course.
Therefore, the standard error of the sample mean is given by
N − n σ2
SE X
= ( ) Var X
= ( ) N −1 n
After understanding the shape of the sampling distribution when the population
is normal and population variance is known. Now you have curiosity to know
what the sampling distribution will be when variance is unknown whether it is
normal or not. Let us discuss the same in the next sub-section.
( )
E X = μ= 96 55
Block 1 Sampling Distributions
Since the standard error is the standard deviation of the sampling distribution
of mean, therefore,
SE
= ( )
X SD
= X 3 ( )
Now, we can answer part (iii) by computing the probability that a randomly
chosen group of 16 batteries has an average life greater than 100 months, that
is
P X > 100
Since the average life of the batteries is normally distributed with a mean of 96
months and a standard deviation of 3 months, therefore, we use the normal
distribution and convert the sample average life to the standard normal Z-
score (as discussed in Unit 1) as follows:
=Z =
( )
X − E X X − 96
SD X ( ) 3
Therefore, the probability that the sample average life of the batteries will be
greater than 100 is 0.0918.
What do we conclude? This probability is low. We conclude that there would
be only a 9.18% (small) chance that the average life of the battery is greater
than 100 months if we selected a sample of 16 batteries.
Let us consider part (iv).
In this case, the population standard deviation is not known, therefore, the
distribution of the sample average life of the batteries follows the t-distribution
with n – 1 degrees of freedom instead of normal.
The mean of the sampling distribution of the average life of the batteries will
be the same as the life of the battery. Therefore, the mean of the sampling
distribution is
( )
E X = μ= 96
We can compute the standard deviation and standard error when the
population standard deviation is not known as
S 13
56
SD =
X ( ) =
n
= 3.25
16
Unit 2 Sampling Distributions of Sample Means
SE
= ( )
X SD
= X 3.25 ( )
So we conclude that the standard error is slightly greater than the previous
case when the population standard deviation is known.
Since the average life of the batteries now follows the t-distribution with n –1
degrees of freedom, therefore, we can convert it into the t-distribution as
follows:
=t
X −E X
=
( )
X − 96
SD X ( )
3.25
To find the value of the above expression, we will use the t-table (Table V)
(given in the Appendix of this volume) because the distribution of the statistic t
follows the t-distribution.
The main difference between the standard normal distribution table and the
t-table is that the body of the standard normal distribution table represents the
probability corresponding to the value of Z whereas the body of the t-table
represents the value or point corresponding to the probability. Here, we have
to find the probability beyond the point 1.23 at 16 –1 = 15 degrees of freedom.
First, we check whether this value lies in the row corresponding to 15 degrees We discuss more about the
of freedom or not. If it lies in the row for df 15 then we read the corresponding t-table in Unit 5.
probability value (α) from the corresponding column heading. If it does not lie
in the row for 15 df then we read the probability value (α) from the column
heading corresponding to just greater and just smaller than 1.23. Thus, from
the t-table, we get
0.10 and P [ t > 0.866] =
P [ t > 1.341] = 0.20
Therefore, we conclude that there are 10% to 20% chance that the average
life of the batteries is greater than 100 months if we select a sample of 16
batteries when the population variance is unknown.
Note: With the help of the t-table, we cannot calculate the exact probability.
But computer packages or software such as R, SPSS, SAS, MINITAB,
R-code
STATA, EXCEL, etc. help us to calculate it exactly. From R, we find it as
pt(q= 1.23, df=15,
P X > 100 =P [ t > 1.23] lower.tail = FALSE)
= 0.11882 = 0.11882
Hence, we conclude that the probability of the average life of the battery is
greater than 100 months if we selected a sample of 16 batteries when the
population variance is unknown and has been increased. 57
Block 1 Sampling Distributions
After the above discussion, you understand that the population standard
deviation (known or unknown) plays an important role in deciding the shape of
the sampling distribution of mean. Now, you can try the following Self
Assessment Question which will make you more user-friendly with the concept
of the sampling distribution of mean when the population is normal with known
or unknown variance.
SAQ 1
Suppose the length of time that a caller is placed on hold when telephoning a
customer service centre is normally distributed with a mean of 40 seconds and
a standard deviation of 5 seconds. If a researcher monitors 20 calls, then
(i) What is the sampling distribution of mean length of hold time?
(ii) Find the mean, standard deviation and standard error of the sampling
distribution.
(iii) Find the probability that the mean length of time on hold in a sample of 20
calls will be within 2 seconds of the population mean.
After understanding the shape of the sampling distribution when the population
is normal and the population variance is known or unknown. I think now you
are interested to know what the sampling distribution will be when the
population is not normal. Let us discuss the same in the next section.
Teaching Experience
Faculty
(in years)
F1 5
F2 6
F3 5
F4 5
F5 5
F6 6
F7 6
F8 12
F9 12
F10 15
That is, We first plot the graph of the population (teaching experience) to know
the form of the population as follows in Fig. 2.3.
5
4
Frequency
0
5 6 12 15 16
Teaching Experiance
From the above figure, we observe that the teaching experience of the faculty
members does not follow a normal distribution. It is right skewed. To know
what the shape of the sampling distribution of the mean will be, let us consider
all possible samples with replacement of size n = 2 from the above population
of teaching experience. In this case, there are Nn = 102 = 100 possible simple
random samples with replacement of size 2. For each sample, we calculate
the sample mean and prepare the sampling distribution as follows in Table
2.4:
Table 2.4: Sampling Distribution of Sample Means
S. No. X Frequency(f)
1 5.0 16
2 5.5 24
3 6.0 9
4 8.5 16
59
Block 1 Sampling Distributions
5 9.0 12
6 10.0 8
7 10.5 6
8 12.0 4
9 13.5 4
10 15.0 1
Total 100
We plot the sampling distribution of mean in Fig 2.4(a). From the figure, you
can notice that for a small sample size n = 2, the shape of the sampling
distribution of mean does not follow a standard distribution (is not normal).
Fig. 2.4: Sampling distribution of mean when population is not normal for n = 2, 3, 5, 30.
Solution: Since the researcher does not know the distribution of the pH level
in that region. However, the researcher collected samples from 35 lakes,
which is greater than 30. Since the sample size is sufficiently large so we can
apply the central limit theorem to find the shape of the sampling distribution of
mean. According to the theorem, it is normally distributed.
The mean of the sampling distribution of mean will be the same as the mean
of the pH (acidity) level in that region. Therefore, the mean of the sampling
distribution of mean is
( )
E X = μ= 6.4
We can compute the standard error when the population standard deviation is
not known as
S
SE
= ( )
X SD
= X ( ) n
0.72 0.72
= = = 0.12
35 5.92
Now, you can try Self Assessment Questions for your practice.
SAQ 2
Consider Example 2 and compute the probability that the sample mean pH
(acidity) level of the water is below 6.4 for the water samples of 35 lakes
collected by the researcher. Also, conclude the results.
We hope that you understood the shape of the sampling distribution of mean
under different situations. We now study the shape of the sampling distribution
of difference of two means. Let us discuss the same in the next section.
Suppose Population-I has mean μ1 and variance σ12 and Population-II has
2
mean μ2 and variance σ 2 .
Population-I Population-II
Mean- Mean- μ2
f μ1
2
Variance- σ 2
2
Variance- σ1
As we discussed in the case of a single sample the sampling distribution X If X and Y are independent
( ) ( ) ( )
E X − Y = E X − E Y = µ1 − µ2
65
Block 1 Sampling Distributions
and variance
(
Var X −= )
Y Var X + Var Y ( ) ( )
σ12 σ 22
= +
n1 n2
(
SE X −=
Y ) Var X − Y ( )
σ12 σ 22
= +
n1 n2
Z=
( X − Y ) − (μ 1 − μ2 )
~ N ( 0,1)
σ12 σ 22
+
n1 n2
t=
( X − Y ) − (μ 1 − μ2 )
S12 S22
+
n1 n2
sample sizes n1 and n2 are large (≥ 30), then according to the central limit
theorem, the sampling distribution of (X − Y) is very closely normally
S2 S2
distributed with mean (µ1− µ2) and variance 1 + 2 whatever the form of
n1 n2
the populations.
Let us move to the next case.
When Population Variances are Unknown and Equal
When the populations from which we draw independent samples are normally
( )
distributed, but their variances are equal σ12 = σ 22 = σ 2 though we do not have
their values, then the sampling distribution of the difference of two means is
again bell-shaped but not normally distributed. In this case, the unknown but
equal population variances σ12 = σ 22 = σ 2 is estimated by pooled sample
variance Sp2 to have a better estimate of the common variance where,
1 n n
=Sp2
n=
∑ 2 2
(Xi − X) + (Yi − Y)
1 + n2 − 2
i 1 =i 1
∑
( )
If sample variances S12 and S22 are known, then we can write the above
expression of pooled sample variance as
1
=Sp2 ( n1 − 1) S12 + ( n2 − 1) S22
n1 + n2 − 2
t=
( X − Y ) − (μ − μ ) ~ t
1 2
(n1 + n2 − 2)
1 1
Sp +
n1 n2
Before describing the next case, let us understand the concept using an
example.
(ii) The probability that the difference between the mean heights of the
players of the samples lies between 2 cm to 5 cm.
• Sample size.
Here, the players are selected from different teams, so the samples are
independent. Here it is given that the heights of the players of both teams are
normally distributed, therefore, the sampling distribution of differences in mean
heights follows either normal or t-distribution depending on whether standard
deviations are known or unknown. Since in this case, the standard deviations
are known, therefore, it is approximately normal whatever the sample sizes.
If X and Y represent the mean heights of the selected players in both teams,
respectively, then we can obtain the mean of the sampling distribution as
follows:
( )
SE X − Y= SD X − Y ( )
σ12 σ 22
= +
n1 n2
25 25
= + = 2.24
10 10
Therefore, by subtracting 2 from each term and then dividing each term by
2.24 in the above inequality, we calculate the required probability as
1 − 2 (X − Y) − 2 5 − 2
P < < = P [ −0.45 < Z < 1.34]
2.24 2.24 2.24
and variance
σ12 σ 22 100 49
Var(X − Y) = + = + = 3.9
n1 n2 40 35
(
SE X − Y = ) (
Var X − Y = ) 3.9= 1.97
Therefore, by subtracting 2 from each term and then dividing each term by
1.97 in the above inequality, we get the required probability as
(X − Y) − 5 6 − 5
P ≥ =P [ Z ≥ 0.51] =
1 − P [ Z < 0.51]
1.97 1.97
1 − 0.6950 =
= 0.305
Thus, we conclude that it would be only a 30.5% chance that the mean scores
of the learners of the first study centre is at least 6 more than the mean scores
70 of the learners of the first study centre.
Unit 2 Sampling Distributions of Sample Means
Now, you can try the following Self Assessment Question before moving to the
next section.
SAQ 3
Continuing with our example of the final exam scores of the learners in MST-
016 (Example 4). What will be the sampling distribution of difference of two
means in the following cases:
(i) If the standard deviations of final exam scores of the learners in the paper
of the course MST-016 of both study centres are unknown.
(ii) If the course coordinator of MST-016 selects 20 learners randomly from
the first study centre and 10 from the second instead of 40 and 35.
example, in the case of assessing the effect of a diet plan to reduce the
weight, we can take weight before and after the diet and subtract the weight
after the diet from the before weight. The difference makes sense too! It is the
weight lost on the diet. Therefore, to obtain the sampling distribution of
difference of two means when the populations are paired, we calculate the
difference of each pair. This approach essentially transforms the paired two-
population data into a one-population, and we can use the same approach
which is used in the sampling distribution of a single mean which is discussed
in Section. 2.2. That is, we draw all possible simple random samples with
replacement of the same size from the population of differences and prepare
the sampling distribution. To know the shape of the sampling distribution of the
paired sample, we follow the different cases as discussed in the sampling
distribution of a single mean.
2.7.1 When Population of Differences is Normally
Distributed
If the population of differences is normally distributed, then the sampling
distribution of mean is also bell-shaped for any sample size. However, the
exact shape of the sampling distribution of mean also depends on whether the
variance/standard deviation of the population of differences is known or
unknown. Therefore, we consider both cases:
When Population Variance of Differences is Known
In this case, when the population of differences is normally distributed, and the
variance of the population is known then the sampling distribution of the
difference of means follows normal distribution regardless of the sample size
n.
In symbolical form, if μD and σD2 denote the mean and variance of the
population of difference and XD denotes the mean of the differences of the
samples of size n drawn from the normal population, then XD will follow a
2
σD
normal distribution with mean μD and variance as we discussed in the
n
sampling distribution of the single mean. Therefore, in symbolic form
σ2
XD ~ N μD , D
n
Here, we use D in the suffix to indicate the difference of population before and
after which differentiate it from the single mean. Also, the quantity
XD − μD
Z= ~ N ( 0,1)
σD
n
follows the standard normal distribution.
When Population Variance of Differences is Unknown
In real-world situations, the variance of the population of differences is not. In
72 such situations, when the population of differences of paired data is normally
Unit 2 Sampling Distributions of Sample Means
distributed, and the variance of the population is unknown then the sampling
distribution of the difference of means is bell-shaped but not normally
( )
distributed. In such situations, we use sample variance SD2 in place of
( ) 2
population variance σ . Due to this replacement, the shape of the sampling
D
distribution of the difference of two means slightly changes from the normal
distribution. Therefore, the quantity
XD − μD
t= ~ t (n−1)
SD
n
follows the t-distribution with n – 1 degrees of freedom.
When the sample size is large (n ≥ 30) then the sampling distribution of
difference of two means for paired data converges to normal distribution
whatever the form of the population may be normal or non-normal
distribution.
Note: Since the sampling distribution is the basis of the statistical inference,
therefore, if we have to draw inferences for the population on the basis of a
small sample (n < 30) then the shape of the population of differences should
be normal. If it is not normal or we do not know the form of the population then
large samples (n1, n2 ≥ 30) will be required otherwise we apply the non-
parametric techniques to draw the inferences. We will discuss the non-
parametric techniques in the third semester.
(i) What will be the sampling distribution of the difference of two means?
(ii) Find the mean, standard deviation and standard error of the sampling
distribution.
(iii) What is the chance that the cholesterol level of the 11 patients is reduced
by at least 32 mg?
( )
E XD= μ=
D 35
SE
= ( )
XD SD
= ( )
XD 1.51
Now, we can answer part (iii) by computing the probability that a randomly
selected group of patients has reduced on average more than 32 mg
cholesterol level using the new diet plan, that is, P XD > 32
=Z =
( )
XD − E XD XD − 35
SD XD( ) 1.51
XD − 35 32 − 35
P XD > 32
= P 1.51 > 1.51= P [ Z > −1.99
= ] P [ Z ≤ 1.99]
From the standard normal Table (Table IV) given in the appendix of this
volume, we get
Therefore, we conclude that it would be 97.69% chance that the new diet plan
would reduce the 32 mg cholesterol level of the patients.
Now, you can try the following Self Assessment Question. 75
Block 1 Sampling Distributions
SAQ 4
Continuing with our example of a new diet plan for lowering the cholesterol
level (Example 5). If the population standard deviation is not known and the
researcher obtained a sample standard deviation of 4 mg, then how will this
affect the sampling distribution and standard error? Do you think this will affect
the probability? Justify your answer.
2.8 SUMMARY
In this unit, we have discussed the following points:
• If the population from which we draw the samples is normally distributed
with population variance/standard deviation is known, then the sampling
distribution of mean will follow a normally distribution and if it is unknown
then the sampling distribution of mean will follow a t-distribution with n –
1 degrees of freedom, regardless the size of the samples.
• If the population from which we draw the samples is not normally
distributed or the shape is unknown, then the shape of the sampling
distribution generally does not specify or does not follow a standard
distribution when the sample size is small. However, if the sample size is
large (30 or more), then the sampling distribution of mean will be
normally distributed, regardless of the shape of the population.
• Two sets of observations are called independent when the observations
are taken independently from two different groups.
• Two sets of observations are called paired when the observations are
taken on the same subject at two different times.
• The sampling distribution of difference of two means when the
populations are independent and normally distributed is also normally
distributed when population variances are known, however, if they are
unknown then the sampling distribution follows the t-distribution.
• The sampling distribution of difference of two means when the
populations are paired and normally distributed is also normally
distributed when population variances are known, however, if it is
unknown then the sampling distribution follows t-distribution.
( )
E X = μ= 40
SE
= ( )
X SD
= ( )
X 1.12
Now, we can answer part (iii) by computing the probability that the mean
length of time on hold in a sample of 20 calls will be within 2 seconds of
the population mean, that is, P 40 − 2 ≤ X ≤ 40 + 2=
P 38 ≤ X ≤ 42
=Z
X −E X
=
( )
X − 40
SD X( )1.12
38 − 40 X − 40 42 − 40
P 38 ≤ X ≤ 42 = P ≤ ≤ = P [ −1.79 ≤ Z ≤ 1.79]
1.12 1.12 1.12
= P [ Z ≤ 1.79] − P [ Z ≤ −1.79]
From the standard normal tables (Tables III and IV) given in the Appendix
of this volume, we get
P 38 ≤ X ≤ 42
= P [ Z ≤ 1.79] − P [ Z ≤ −1.79]
Therefore, we conclude that there would be 92.66% chance that the mean
length of time on hold in a sample of 20 calls will be within 2 seconds of
the population mean.
2. We have to calculate the probability that the sample mean pH level of the
water is below 6.4, that is, P X < 6.5 .
=Z
X −E X
=
( )
X − 6.4
SD X( )0.12
From the standard normal table (Table III) given in the Appendix of this
volume, we get
P X < −2.5 =0.0062
(ii) Since it is given that the distribution of test scores of the first study centre
is left-skewed and the sample size is small (< 30), therefore, the sampling
78 distribution of difference of two mean scores does not follow the
Unit 2 Sampling Distributions of Sample Means
t-distribution because for the t-distribution both populations should be
normally distributed whatever may be the sample sizes. For the two distinct
Since the distribution of test scores of the first study centre is left skewed populations:
and the sample size is small (< 30), therefore, the sampling distribution • If the sample sizes are
difference of two mean scores does not follow the normal distribution small, the shape of
because for the normal distribution either both the populations should be distributions are
normal or sample sizes large. important (should be
4. Since the differences of the cholesterol levels before and after the new normal)
diet plan follow a normal distribution and the standard deviation is not • If the sample sizes are
known, therefore, the sampling distribution of the difference of two means large, the shape
follows the t-distribution with n – 1 degrees of freedom instead of normal. distributions are not
We can compute the standard error when the population standard important (need not be
SD 4
( )
SE X=
D SD X=
D ( ) =
n
= 1.21
11
So, we conclude that the standard error is slightly smaller than in the
previous case.
We can compute the probability that a randomly selected group of
patients has reduced on an average more than 32 mg cholesterol level
using the new diet plan, that is, P XD > 32
Since the mean cholesterol level follows the t-distribution instead of the
normal distribution, therefore, we can convert it into t-distribution as
follows:
= t =
XD − E XD ( )
XD − 35
SD XD ( )1.21
Therefore, we conclude that there is 97.5% to 99% chance that the new
diet plan reduces the 32 mg cholesterol level of the patients. 79
Block 1 Sampling Distributions
Note: Here, we cannot calculate the exact probability using the t-table.
But with the help of computer packages or software such as R, SPSS,
SAS, MINITAB, STATA, EXCEL, etc. we can calculate it exactly.
80
UNIT 3
SAMPLING DISTRIBUTIONS OF
SAMPLE PROPORTIONS
AND VARIANCES
Structure
3.1 Introduction 3.5 Sampling Distribution of Ratio
of Two Sample Variances
Expected Learning Outcomes
When Populations are Normally
3.2 Sampling Distribution of
Distributed
Sample Proportion
When Populations are not Normally
3.3 Sampling Distribution of
Distributed
Difference of Two Sample
Proportions 3.6 Summary
3.4 Sampling Distribution of 3.7 Terminal Questions
Sample Variance
3.8 Solutions/Answers
When Population is Normally
Distributed
unit, we discussed the concept of the sampling distribution of mean and the appropriate Unit or
section before proceeding:
difference of two sample means. However, sometimes, we deal with the data
collected in the form of counts or the collected data classified into two • Basic Concepts of
categories or groups according to an attribute. For example, how many people Sampling Distribution
in a sample choose Thums up as their soft drink, how many people are (Unit 1).
educated in a city, how many patients recovered from COVID-19, etc. In such • Sampling
situations, we use sample proportion instead of mean and the sampling Distributions of
Means (Unit 2).
distribution of sample proportion is a fundamental concept in statistics that plays
a pivotal role in making precise inferences about population proportion from
sample data. 81
Unit Writer- Dr. Prabhat Kumar Sangal, School of Sciences, IGNOU, New Delhi
Block 1 Sampling Distributions
Similarly, in many practical situations, we have also been concerned with
variability. Especially, when a little variance in diameter as possible is crucial
when manufacturing things that fit together, like pipes and ball bearings. If such
products do not fit together correctly then we discard such products. In such
situations, it is important to study the variability of the population and the
sampling distribution of sample variance plays a pivotal role in making precise
inferences about population variance from sample data.
For example, if there are 400 learners enrolled in the MSCAST programme and
out of these 250 learners successfully complete the programme then we can
compute the proportion of pass learners as
Number of the learners complete the programme 250
=p = = 0.625
total learners enroll in the programme 400
You have already tasted the flavour of the sampling distribution of mean with
the help of examples in Unit 2. To study the shape of the sampling distribution
of the sample proportion, we proceed in the same way that we have followed in
the case of the sampling distribution of the sample mean. That is, we draw all
possible samples of the same size from the population and calculate the sample
proportion for each sample instead of the sample mean. After calculating the
sample proportion for each sample, we construct the sampling distribution of
sample proportions. The distribution will be obtained is known as the sampling
The sampling distribution
distribution of sample proportion. Therefore, we can define it as:
of the sample proportion is
a theoretical probability
“The probability distribution of all possible values of the sample
distribution of sample
proportion that would be obtained by drawing all possible samples of the
proportions that would be
same size from the population is called the sampling distribution of
obtained by drawing all
sample proportion or simply says sampling distribution of proportion.”
possible samples of the
Now, the question may arise in your mind “What is the shape of the sampling same size from the
distribution of sample proportion, is it normal, or t?” as we have obtained in the population.
case of the sampling distribution of sample mean. But there is a main difference
between mean and proportion. The mean is used when we deal with continuous
variables such as height, income, etc. But in proportion, we will deal with data 83
Block 1 Sampling Distributions
collected in the form of counts when the collected data is classified into two
categories or groups according to an attribute. We know that (Unit 10 of MST-
012: Probability and Probability Distribution), if our data is classified into two
categories or groups (success or failure) according to an attribute and we count
how many of them possess a given attribute, then the data follow a binomial
distribution. It means that the form of the population is binomial instead of
normal in the case of the sampling distribution of sample proportion. Let us see
what the shape of the sampling distribution of sample proportion will be when
the population is binomial. For a better understanding, we consider an example
in which the size of the population is very small.
Consider the example of our cute children of play school discussed in Unit 2
and consider only the first five children out of a group of 9 so that we can easily
show all possible samples. Suppose the play-school teacher is now interested
to know the proportion of children who like to dance. To assess this, she asked
each child whether he/she liked to dance or not and obtained the information as
follows:
6
Number of children
1
Like Dislike
Like Dance
Fig. 3.1: Distribution of the number of children who like and dislike dance.
From the above figure, we observe that the number of children who like to
dance does not follow a normal distribution while it follows a binomial
distribution.
Let us prepare the sampling distribution of sample proportion. To obtain this, let
us consider all possible samples with replacement of size n = 2 from the above
population. There are Nn = 62 = 36 possible simple random samples with
replacement of size 2 and the proportions of children who like to dance of each
84 sample are given in the following table.
Unit 3 Sampling Distributions of Sample Proportions and Variances
Sample Children Name Sample Sample Sample Children Name Sample Sample
Observations Proportion(p) Observations Proportion(p)
2 (Avaya, Ishaan) (Yes, No) 0.5 20 (Zara, Ishaan) (Yes, No) 0.5
5 (Avaya, Kavya) (Yes, No) 0.5 23 (Zara, Kavya) (Yes, No) 0.5
7 (Ishaan, Avaya) (No, Yes) 0.5 25 (Kavya, Avaya) (No, Yes) 0.5
9 (Ishaan, Ria) (No, Yes) 0.5 27 (Kavya, Ria) (No, Yes) 0.5
10 (Ishaan, Zara) (No, Yes) 0.5 28 (Kavya, Zara) (No, Yes) 0.5
12 (Ishaan, Parth) (No, Yes) 0.5 30 (Kavya, Parth) (No, Yes) 0.5
14 (Ria, Ishaan) (Yes, No) 0.5 32 (Parth, Ishaan) (Yes, No) 0.5
17 (Ria, Kavya) (Yes, No) 0.5 35 (Parth, Kavya) (Yes, No) 0.5
Total 24
Mean 0.67
From the above table, you can see that the sample proportion varies from
sample to sample. To obtain the distribution of the sample proportions, we
arrange the values of the sample proportion in ascending order and calculate
the frequency corresponding to each value as discussed in Units 1 and 2. The
obtained sampling distribution is shown in Table 3.2.
Table 3.2: Sampling Distribution of Sample Proportion
S. Sample
Frequency Probability
No. Proportion(p)
1 0 4 4/36
2 0.5 16 16/36
3 1 16 16/36
1
= ( 0 × 4 + 0.5 × 16 + 1× 16=) 0.67
= P
36
You see that the mean of the sampling distribution of sample proportions is
equal to the population proportion. Thus, we can say that to find the exact
estimate of the unknown population parameter, we first find the sampling
distribution of sample proportion and then compute the mean of the obtained
sampling distribution of proportion.
We now calculate the standard deviation of the sampling distribution of sample
proportion as
1 k
∑ fi ( pi − p )
2
SE(p)
=
K i=1
1
SE (=
p) 4 × ( 0 − 0.67 ) + 16 × ( 0.5 − 0.67 ) + 16 × (1 − 0.67 )
2 2 2
36
1
= (1.7956 + 0.4626 + 1.7424
= ) 0.3334
36
Fig. 3.2: Sampling distribution of sample proportion for sample size n = 2, 3, 5 and 30.
86
Unit 3 Sampling Distributions of Sample Proportions and Variances
Let us see what happens as we increase the sample size, whether it converges
to normal distribution or not. We also plot the sampling distribution of sample
proportion when sample size n = 3, 5 and 30 in Fig. 3.2(b) to 3.2(d).
From Fig. 3.2, you can observe that as we increase the same size the sampling
distribution of proportion approaches the normal distribution.
After knowing the shape of the sampling distribution of proportion, you may be
interested to know the mean and variance of the sampling distribution. Let us
find the mean, variance and standard error of the sampling distribution of
proportion.
Here, we divide the units of the population into two groups on the basis of an
attribute and the sample observations are independent then the number of units
that possess the given attribute (number of successes) follows a binomial
distribution (Unit 10 of the course MST-012) with mean
E(X) = nP
and variance
= nP (1 − P )
Var(X)
and variance
X 1 Var ( aX ) a2 Var ( X )
Var(p) = Var =
= Var(X)
n n2
1
= nP (1 − P ) ( X ) nP (1 − P )
Var=
n2
P (1 − P )
Var(p) =
n
Also, by the definition of standard error, we can obtain the standard error of The standard deviation of a
sample proportion as sampling distribution of a
P (1 − P ) statistic is known as
SE
= (p ) SD
= (p ) Var(p)
= standard error.
n
A question may arise “Can we apply the central limit theorem for the
sampling distribution of sample proportion?” The answer is “Yes” because
the sample proportion could be thought of as a mean. If we label the success as
1 and the failure as 0, then the sample mean becomes the sample proportion as
n
1 1 + 0 + 0 + 1 + .....0 X
X
=
n
∑ X=i n
= = p
n
i=1
Therefore, we can also apply the central limit theorem for large samples. Since
the number of successes follows binomial distribution and Binomial distribution 87
Block 1 Sampling Distributions
converses to normal distribution when n is large, and the probability of success
(P) remains close to 0.5. Therefore, we require some conditions on both n and
Some authors use the P. The central limit theorem states that the sampling distribution of the sample
following conditions for the proportions converges to the normal distribution if
normality of binomial
• nP > 15 and
distribution
• n(1 – P) > 15
• nP > 5 or > 10
• n(1 – P) > 5 or 10 these conditions are called normality conditions. Thus, we can say that if the
sample size is sufficiently large, such that nP > 15 and n(1 – P) > 15 then by
central limit theorem, the sampling distribution of sample proportion p is
approximately normally distributed with mean P and variance P(1 – P)/n, that is,
P (1 − P )
p ~ N P,
n
In this case, we use the normal distribution to answer the probability questions
about sample proportions and the quantity (Z)
p −P
Z= ~ N ( 0,1)
P (1 − P )
n
• np < 15 or
• n(1 – P) < 15
then the sampling distribution of the number of the units possesses a given
attribute follows a binomial distribution with mean np and variance nP(1 – P),
and we must use the binomial distribution to answer probability questions about
sample proportions by converting the proportion to the number of items with the
characteristic of interest, X (See Example 1).
(i) We can compute the sample proportion as the number of orders (X) that
are shipped within 24 hours divided by the total number of orders (n) in the
sample. Therefore,
X 210
p= = = 0.84
n 250
(ii) The sample proportion p is approximately normally distributed, if nP > 15
and n(1 – P) > 15, Therefore, first, we check these conditions as follows:
nP = 250 × 0.90 = 225 >15 and
Since both conditions are fulfilled, therefore, we can say that the sample
size is large enough to take the normal distribution approximation of the
sampling distribution of sample proportion.
We can find the mean and standard deviation of the sampling distribution of
sample proportion as
E(p)= P= 0.90
SE
= (p ) SD
= (p ) 0.019
89
Block 1 Sampling Distributions
(iii) The sample proportion computed from a sample of size 250 will be within 5
percentage points of the true population proportion to lie between 0.90 –
0.05 = 0.85 and 0.90 + 0.05 = 0.95. Therefore, the required probability will
be P [0.85 < p < 0.95] .
Hence, we can conclude that there is 99% chance that the sample
proportion computed from a sample of size 250 will be within 5 percentage
points of the true population proportion.
(iv) Since in this case, the sample size n is 20, so we check whether the
normality conditions are fulfilled or not, that is, nP > 15 and n(1 – P) > 15.
Here,
nP = 20 × 0.90 = 1.8 < 15
n (1 − P ) = 20 × (1 − 0.90)= 2 < 15
Since both are less than 15, therefore, the sampling distribution of
proportion will follow the binomial distribution instead of normal.
We now compute the required probability P [0.85 < p < 0.95] using the
binomial distribution.
To calculate the required probability, we use the binomial distribution, that
is, we have to convert 0.85 and 0.95 into the number of orders by
multiplying these with the sample size, that is, n = 20.
Therefore, we can transform the required probability into a binomial as
P[X ] Cx p x (1 − p )
n n− x
= x=
P[X =
18] = C18 ( 0.90 ) (1 − 0.90 )
20 18 20 −18
0.2852
=
Hence, we can conclude that there is only 28.52% chance that the sample
proportion computed from a sample of size 20 will be within 5 percentage
points of the true population proportion.
Now, it is time for you to try the following Self Assessment Question to make
sure that you have learnt the sampling distribution of sample proportion.
SAQ 1
A machine produces a large number of items of which 8% are found to be
defective. If 40 items are selected randomly from the production, then
(i) What will be the sampling distribution of sample proportion?
(ii) Find the mean, standard deviation and standard error of the sampling
distribution.
(iii) Calculate the probability that less than or equal to 12% defectives are found
in the sample.
P (1 − P1 ) P2 (1 − P2 )
p1 ~ N P1, 1 and p2 ~ N P2 ,
n1 n2
Also, by the property of normal distribution described in Unit 14 of MST 012, the
sampling distribution of the difference of two proportions also follows a normal
distribution with mean
E ( p1 − p2 ) =E ( p1 ) − E ( p2 ) =P1 − P2
and variance
P1 (1 − P1 ) P2 (1 − P2 )
Var ( p1 − p=
2) Var ( p1 ) − Var ( p=
2) +
n1 n2
That is,
P (1 − P1 ) P2 (1 − P2 )
p1 − p2 ~ N P1 − P2 , 1 +
n1 n2
P1 (1 − P1 ) P2 (1 − P2 )
SE ( p1 −=
p2 ) Var ( p1 − =
p2 ) +
n1 n2
92 Therefore, we can say that the sampling distribution of difference of two sample
Unit 3 Sampling Distributions of Sample Proportions and Variances
Z=
(p1 − p2 ) − (P1 − P2 ) ~ N ( 0,1)
P1 (1 − P1 ) P2 (1 − P2 )
+
n1 n2
• n2P2 ≤ 15 or n2 (1 − P2 ) ≤ 15
then the sampling distribution of the difference of two proportions may not be
approximately normal. For smaller sample sizes, this approximation may not
hold, and the distribution could be skewed or has heavier tails and does not
follow standard form.
I think you understand the shape of the sampling distribution of difference of two
proportions in different cases. We also explain various forms of the sampling
distribution under different conditions using the flow chart given in Fig. 3.4 which
helps you to quickly judge the shape of the sampling distribution of difference of
two proportions.
Since all conditions are satisfied, therefore, we can assume that the sampling
distribution of difference of two sample proportions is approximately normally
distributed.
Thus, we can compute the mean and standard deviation of the sampling
distribution as
E ( p1 − p2 ) = P1 − P2 = 0.30 − 0.20 = 0.10
P1 (1 − P1 ) P2 (1 − P2 )
SD ( p1=
− p2 ) +
n1 n2
By the definition of the standard error, we can compute the standard error as
SE ( p1 − p2=
) SD (p1 − p2=) 0.043
Now, we have to compute the probability that the difference in sample
proportions is less than 0.05. Therefore, we can represent it in symbolic form as
P [p1 − p2 < 0.05]
Therefore, by subtracting 0.10 from each term and then dividing each term by
0.043 in the above inequality, we get the required probability as
(p − p2 ) − 0.10 0.05 − 0.10
P 1 < = P [ Z < −1.16]
0.043 0.043
SAQ 2
In Example 2, if 40 people are selected randomly from each population
independently instead of 200 people then how will this affect the sampling
distribution? Do you think this will affect the probability? Justify your answer.
I hope that you understand the shape of the sampling distributions of sample
proportion and difference of two proportions when the samples are independent
and now you may have the curiosity to know the shape of the sampling
distribution of sample variance. Therefore, we now discuss the sampling
distribution of sample variance in the next session.
• The variation in the weight of the juice packet also plays an important role
in the goodwill of the company.
Child Name Avaya Ishaan Ria Zara Kavya Parth Rayaan Nyra Yash
Score 2 5 4 3 1 3 2 4 3
To know whether the shape of the population (test scores) is normal or not, we
first plot the graph of the scores of the children as shown in Fig. 3.5.
96
Unit 3 Sampling Distributions of Sample Proportions and Variances
Frequency
2
0
1 2 3 4 5
Score (out of 5)
Fig. 3.5: The distribution of the scores of the children in the listening task.
From the above figure, you can observe that the test scores of the children
follow an approximate bell-shaped normal distribution.
To obtain the shape of the sampling distribution of the sample variance when
the sample is taken from a normal population, let us consider all possible
samples with replacement of size n = 2 from the population of the test scores of
the children. There are Nn = 92 = 81 possible simple random samples with
replacement of size 2. All possible samples of size n = 2 are given in the
second column of Table 3.3. We also calculate the variance of each sample.
The obtained sample variances are given in the last column of the same table.
Table 3.3: Samples and Sample Variances
Sample
Sample Sample in Term of Sample
Observations
number Children Variance
(scores)
… … … …
Here, we have given some of these possible samples to reduce the space. You
can prepare the same as we have discussed in Units 1 and 2. To obtain the
distribution of the sample variances, we arrange the values of the sample
variance in ascending order and calculate the frequency corresponding to each
value as discussed in Units 1 and 2. The obtained sampling distribution of
sample variance is shown in Table 3.4. 97
Block 1 Sampling Distributions
Table 3.4: Sampling Distribution of Sample Variance
Sample
S. No. Frequency(f) Probability(p)
Variance
1 0 19 19/81
2 0.25 32 32/81
3 1.00 20 20/81
4 2.25 8 8/81
5 4.00 2 2/81
6 4.50 0 0/81
Total 81 1
Fig. 3.6: Sampling distribution of variance when population is normal for sample size n = 2,
3, 5 and 30.
From Fig. 3.6(a), you can observe the shape of the distribution of sample
variance, and you can notice that for a small sample size n = 2, the shape of the
sampling distributions of sample variance does not follow a normal. However, it
98 is right-skewed.
Unit 3 Sampling Distributions of Sample Proportions and Variances
Let us see what happens as we increase the sample size whether it converges
to normal distribution or not. We plot the sampling distribution of sample
variance when sample size n = 3, 5 and 30 in Fig. 3.6(b) to 3.6(d), respectively.
From Fig. 3.6, you can observe that as we increase the sample size the
sampling distribution approaches normal distribution.
The distribution of sample variance cannot be obtained directly. In 1875,
German statistician Friedrich Robert Helmert made some transformations in the
sample variance and obtained a new distribution which is known as the chi-
square distribution as the distribution of the sample variance for a normal
population. However, the exact shape of the sampling distribution of sample
variance also depends on whether the population mean is known or unknown.
Therefore, we consider both cases as follows:
Population Mean is Known
When we draw the samples from the normal population with known mean µ and
variance σ2, then we calculate sample variance using the following formula:
1 n
∑ ( Xi − μ )
2
=S2
n i=1
∑ ( X − μ)
2
2
2 nS i
χ= = i =1
~ χ(2n)
σ2 σ2
When the population mean is unknown then we use the sample mean X in
place of the population mean (µ). However, due to the discrepancy between the
sample mean and population population mean, we calculate sample variance
using the following formula:
1 n
( )
2
=S2 ∑ Xi − X
n − 1 i =1
− 1) S2 ∑
2
(Xi − X)
=χ2
(n= i =1
~ χ(2n−1)
σ2 σ2
follows the chi-square distribution with (n − 1) degrees of freedom.
After knowing the shape of the sampling distribution of sample variance when
the population is normally distributed and the population mean is known or
unknown, you may be interested to know the mean and variance of the
sampling distribution of sample variance.
In practice, only one random sample is selected, and the concept of the
sampling distribution is used to draw the inference about the population
parameters. If X1, X2, …, Xn is a random sample of size n taken from a normal
population with mean µ and known variance σ2 then the mean and variance of
the sampling distribution of sample variance also depend whether mean of the
population is known or not known. Therefore, we consider the following cases:
When population mean is known
If the population mean is known, then the mean and variance of the sampling
distribution of sample variance are given as follows:
(n − 1) σ2
of S2 E=
Mean= S2 ( ) n
and variance
2 ( n − 1) σ 4
Var(S2 ) =
n2
Similarly, the standard error is given by
2 ( n − 1) σ4
SE
= S( )
2
SD
= S ( ) 2
Var
= S ( ) 2
n2
2σ 4
SE
= ( )
S2 SD
= S2 ( ) Var
= ( )
S2
n −1
The proofs of the above expressions are out of scope.
Example 3: A manufacturer of steel ball bearings has found that the distribution
of the diameter of the steel ball bearings follows a normal distribution with a
variance of 0.18 inches2. If the manufacturer took a random sample of 25 steel
ball bearings, then
(i) What will the sampling distribution of sample variance be?
(ii) Find the mean, standard deviation and standard error of the sampling
distribution.
(iii) What is the probability that a random sample of 25 ball bearings will result
in a sample variance of at least 0.2 inches2?
σ2 = 0.18, n = 25
Since it is given that the diameter of the steel ball bearings follows a normal
distribution with a variance of 0.18 inches2, therefore, the sampling distribution
of
( n − 1) S2
follows the chi-square distribution. Since the population mean is
σ2
unknown, therefore, it follows the chi-square distribution with n – 1 degrees of
freedom.
We can compute the mean and standard deviation when population means is
unknown as
2σ 4 2 × 0.182
( ) S2
E S2 =σ2 =0.18 and SD= ( ) =
n −1
= 0.052
24
Since the standard error is the standard deviation of the sampling distribution,
therefore,
SE
= S2( )
SD
= S2 0.052( )
Now, we can answer part (iii) by computing the probability that the sample
variance is at least 0.2 inches2 as follows:
P S2 ≥ 0.2
To get the value of this probability, we convert variate S2 into the chi-square
variate which follows the chi-square distribution with n – 1 = 24 degrees of
freedom by the transformation
2 (n − 1)S2 24S2
χ
= ( n −1) =
σ2 0.18
Therefore, multiplying each term by 24 and then dividing each term by 0.18 in
the above inequality. We get the required probability as
24S2 24 × 0.2
= P χ( 24 ) ≥ 26.67
P ≥ 2 We discuss more about the
0.18 0.18 chi-square table in Unit 4.
To find the value of the above expression, we will use the chi-square table
(Table VI) given in the Appendix at the end of this volume. The chi-square table
is similar to the t-table in the sense that the body of the chi-square table
101
Block 1 Sampling Distributions
Therefore, the above probability must lie between 0.10 and 0.90, that is,
Therefore, we conclude that there is 10% to 90% chance that the sample
variance is at least 0.2 inches2.
Note: With the help of the chi-square table, we cannot calculate the exact
probability. But computer packages or software such as R, SPSS, SAS,
MINITAB, STATA, EXCEL, etc. help us to calculate it exactly. From R, we find it
as
After reading the above discussion, you can try the following Self Assessment
Question.
SAQ 3
Consider Example 3 of steel ball bearings. If the average diameter of the steel
ball bearings is known to be 2 inches, then how will this affect the sampling
distribution and standard error? Do you think this will affect the probability?
Justify your answer.
• A doctor wants to test whether the variance in the weight of the newly born
baby girl is less than the baby boy,
S12
F= ~ F(n1 −1,n2 −1)
S22
3.6 SUMMARY
In this unit, we have discussed the following points:
• The sampling distribution of the sample proportions converges to the
normal distribution with mean P and variance P(1 – P)/n if nP >15 and 105
Block 1 Sampling Distributions
n(1 – P) > 15 otherwise it follows a binomial distribution.
• The sampling distribution of difference of two sample proportions
converges to the normal distribution with mean P1 − P2 and variance
P1 (1 − P1 ) P2 (1 − P2 )
+ if
n1 n2
If any one of the above conditions is not fulfilled, then the sampling
distribution of difference of two sample proportions does not follow normal.
• The sampling distribution of variance follows a chi-square distribution with n
and n – 1 degrees of freedom when the population follows a normal
distribution and the population mean is known or unknown, respectively. If
the population is not normally distributed, then it does not follow chi-square
as well as normal distribution for small samples.
• The sampling distribution of ratio of sample variances follows the
F-distribution with (n1 – 1, n2 – 1) degrees of freedom when the populations
follow a normal distribution. If the populations are not normally distributed,
then it does not follow the F as well as normal distribution for a small
sample.
(ii) We can compute the mean and standard of the sampling distribution of
sample proportion as
E(p)= P= 0.08
P (1 − P ) 0.08 × 0.92
SD(p)
= = = 0.00184 0.0429
=
n 40
106
Unit 3 Sampling Distributions of Sample Proportions and Variances
Hence, we can conclude that there is only 78.68% chance that the
machine produces 12% defective items.
Since sample sizes have been changed so we have to check whether the
normality conditions are satisfied or not in the current scenario to know the
shape of the sampling distribution of difference of two sample proportions
as
n1P1 = 40 × 0.30 = 12 < 15, n1 (1 − P1 ) = 40 × (1 − 0.30 ) = 28 > 15,
Since n1P=
1 12 < 15 and n2P2= 8 < 15 conditions are not satisfied,
therefore, we cannot assume that the sampling distribution of difference of
two sample proportions is approximately normally distributed. In this
situation, we cannot calculate the required probability because we do not
know the form of the sampling distribution of difference of two sample
proportions.
σ2 = 0.18, n = 25, µ = 2
Since it is given that the diameter of the steel ball bearings follows a normal
distribution with the mean of 2 inches and variance of 0.18 inches2, 107
Block 1 Sampling Distributions
2
nS
therefore, the sampling distribution of follows the chi-square
σ2
distribution with n degrees of freedom because the population mean is
known.
Now, we compute the probability that the sample variance is at least 0.2
inches2, that is,
P S2 ≥ 0.2
To get the value of this probability, we convert variate S2 into the chi-square
variate which follows the chi-square distribution with n = 25 degrees of
freedom by the transformation
n S2 25S2
χ(2n=
) =
σ2 0.18
Therefore, multiplying each term by 25 and then dividing each term by 0.18
in the above inequality. We get the required probability as
25S2 25 × 0.2
= P χ( 25 ) ≥ 27.78
2
P ≥
0.18 0.18
To find the value of the above expression, we will use the chi-square table
(Table VI) given in the Appendix at the end of this volume as discussed in
Example 3. Here, we have to find the probability beyond the point 27.78 at
25 degrees of freedom. So, we check whether this value lies in the row
corresponding to 25 degrees of freedom or not. Since it is not there so we
read the probability value (α) from the column heading corresponding to
just greater and just smaller than 27.78. Thus, from the chi-square table,
we get
P χ(225 ) ≥ 16.47 =
0.90 and P χ(225 ) ≥ 34.38 =
0.10.
Therefore, the above probability must lie between 0.10 and 0.90, that is,
0.10 < P χ2 ≥ 27.78
= ( )
P S2 ≥ 0.2 < 0.90
Thus, we conclude that there is 10% to 90% chance that the sample
variance is at least 0.2 inches2.
Terminal Questions (TQs)
To check whether the shape of the sampling distribution of difference of two
proportions follows the normal distribution or not, we check whether the
following conditions are satisfied or not. Therefore,
nP = 200 × 0.08 = 16 > 15
Since both are greater than 15, therefore, we can assume that the sampling
distribution of sample proportion is approximately normally distributed with
mean nP and variance P(1 – P)/n.
108
UNIT 4
SAMPLING DISTRIBUTIONS
ASSOCIATED WITH
NORMAL POPULATIONS-I
Structure
4.1 Introduction 4.7 t-distribution
Expected Learning Outcomes Probability Density Curve of t-
distribution
4.2 Concept of Degrees of
Freedom Summary Measures of t-distribution
some basic definitions related to it. In Units 2 and 3, we have discussed the knowledge of any of these
terms, you should review
sampling distributions of various sample statistics such as sample mean,
the appropriate Unit or
sample proportion, sample variance, etc. You have observed that the
section before proceeding:
sampling distribution of mean follows normal or t-distributions under different
conditions whereas the sampling distribution of variance and ratio of two • Sampling
variances follow chi-square and F-distributions, respectively. You have studied Distributions for
the normal/ standard normal distribution in detail in the MST-012: Probability Means and Variance
(Units 2 and 3).
and Probability Distributions course. Therefore, we will focus only on other
important sampling distributions. In this unit, we will discuss the chi-square
109
Unit Writer- Dr. Prabhat Kumar Sangal, School of Sciences, IGNOU, New Delhi
Block 1 Sampling Distributions
and t-sampling distribution.
This unit is divided into 13 sections. Section 4.1 is introductive in nature. The
chi-square, t and F distributions are described with the help of the degrees of
freedom, therefore, in Section 4.2, the concept of degrees of freedom is
described in detail. The chi-square distribution with its probability curve,
summary measures, relation to other distributions, properties and applications
are discussed in Sections 4.3 to 4.6. Similarly, the t-distribution with its
probability curve, summary measures, relation to other distributions, properties
and applications are discussed in Sections 4.7 to 4.10. The unit ends by
providing a summary of what we have discussed in this unit in Section 4.11.
The terminal questions and the solution of the SAQs/TQs are given in
Sections 4.12 and 4.13, respectively.
In the next unit, we shall discuss the F-distribution and how to obtain the
tabulated values of the chi-square, t and F-distributions.
Here, it is noted that the sum of squares of deviations taken from the sample
mean is divided by (n – 1) instead of sample size n. The rationale relates to
the deviations' degrees of freedom.
Firstly, it should be noted that the deviations are calculated from the sample
mean rather than the population mean. It is so because the population mean is
generally not known. When we do not know the population mean then we can
take the deviation from any number as we choose. But the best number to
choose is the sample mean because it will minimise the sum of squares of the
deviations. Therefore, we have imposed one restriction so we lose one degree 111
Block 1 Sampling Distributions
of freedom. Also, when we take the deviation from the sample mean then it will
be smaller than the deviation from the population mean. Therefore, we divide
the sum of squared deviations by its degrees of freedom n –1 instead of n to
compensate for the downward bias. Here, we can also say that we estimate
one parameter (population mean) by sample data (sample mean) so we lose
one degree of freedom. Hence, in statistical inference, we can also define the
degrees of freedom as
The degrees of freedom is the total number of observations in the
sample minus the number of estimated parameters.
A more complicated instance of degrees of freedom arises when employing an
analysis of variance (ANOVA) approach as covered in the MST-013: Survey
Sampling and Design of Experiment-I course. We now try to understand the
concept of degrees of freedom in the context of analysis of variance as used in
MST-013.
In one-way ANOVA, we consider three quantities:
• the total sum of squares (TSS),
• the sum of squares due to treatment (SST), and
• the sum of squares due to error (SSE).
We discuss degrees of freedom for each one by one. The formula of
computing TSS is
k ni
∑∑ ( X )
2
TSS
= ij −X
=i 1=j 1
It is the sum of the squares of the deviations of all data points (X11 + X12 + …+
Xkn j ) from the grand mean. Here, we also impose one restriction, that is, we
take the deviation from the grand mean instead of any number or we use an
estimate of the population mean so we lose one degree of freedom. Therefore,
the degrees of freedom for TSS is n –1 instead of n.
We now come to the second quantity, that is, the sum of squares due to
treatment (SST). The formula for computing SST is
k
∑ ( X − X)
2
SST
= i
i =1
It is the sum of squares of the deviations of all sample means from the grand
mean. Here, we can consider k sample means as k independent data points,
and the grand mean as the mean of all these sample means so again one
restriction of taking a deviation from the mean of sample means instead of the
population mean so it reduces again one degree of freedom. Therefore, the
degrees of freedom for SST is k – 1 instead of k.
We now consider the sum of squares due to error (SSE). The formula for
computing SSE is
k ni
∑∑ ( X )
2
SSE
= ij − Xi
=i 1=j 1
one restriction that the deviation is taken from its sample mean so each group
has ni – 1 degrees of freedom. There are k groups so the total degrees for
SSE is n1 – 1 + n1 – 1 + …+ nk – 1 = n – k.
Let us discuss the degrees of freedom with the help of an example.
Example 1: Let us suppose you received a summary statement of your credit
card from your bank as follows:
Transition Description Amount
Date
05/11/2023 INSTL 1/12 iPHONE X POWER 5800.00
12/11/2023 SM NORTH DEPT STORE 8000.00
15/11/2023 SM NORTH TRAVEL CLUB QUEZON CI 3200.00
20/11/2023 SM NORTH HYPER MARKET QUEZON CIT 6500.00
25/11/2023 INSTL 2/6 ELECTRO WORLD 2500.00
Suppose you want to calculate the sum of squares of deviations (SSD) of the
amount paid about a point so that it will be minimum.
(i) What number should you choose to minimise the SSD?
(ii) Calculate the SSD with the chosen number.
(iii) What is the degrees of freedom for the calculated SSD? Also, calculate
the sample standard deviation (SD).
Solution: As we know the sum of squares of deviation (SSD) is minimum
when we take deviations from the mean, therefore, the statement mean
n
∑ ( X − X)
2
minimises the SSD. We now calculate SSD, that is, i as follows:
i =1
n
Calculation for ∑ ( Xi − X )
2
i=1
∑(X )
2
∑X i = 28500 i −X = 12680000
i=1 i=1
∑ ( X − X)
2
For calculating i , first of all, we have to calculate mean (X) ,
i =1
therefore,
1 n 28500
=X = ∑
n i=1
Xi = 5700
5
Therefore, we can calculate the SSD of the data as
n
∑ ( X − X)
2
i 12680000
=
i =1
∑(X )
2
−X
i
12680000
Sample SD = =
i =1
= 1780.449
df 4
I think you have understood the concept of the degrees of freedom and how to
calculate it. Therefore, try the following Self Assessment Question to assess
your understanding of the concept of degrees of freedom.
SAQ 1
A professor of statistics selected some students of the MSCAST programme
from three study centres (SCs) of IGNOU and noted their marks in the MST-
016 paper. The obtained marks (out of 100) are given as follows:
SC I SC II SC III
70 58 69
94 75 73
67 51 62
82 69 68
87 52
92
(i) If you can take the deviations of these data from three numbers as you
select, and you want to minimize the sum of the squared deviations
(SSD), what numbers would you select?
(ii) What is the minimised SSD?
(iii) How many degrees of freedom are connected to this SSD?
(iv) Compute the mean squared deviation (MSD) by dividing the SSD by its
degrees of freedom.
After understanding the concept of degrees of freedom. Let us take the chi-
square distribution in the next section.
=χ 2 (n − 1) S2 ~ χ(2n−1)
2 (1857 – 1936)
σ
Karl Pearson is the father
follows the chi-square distribution with (n −1) degrees of freedom instead of n of modern statistics. He
because it is based on S2 which has (n −1) degrees of freedom. founded the first statistics
department in the world at
Let us take an example to understand how to calculate the chi-square statistic
University College
and its degrees of freedom.
London.
Example 2: A battery company has developed a new laptop battery and
believes that the battery lasts, on average, 12 hours on a single charge with a
standard deviation of 3 hours. The manufacturing division of the company
performs a quality control test on it. They randomly select 10 batteries, and it
is noted that the mean and standard variation of the selected batteries are 11
hours and 4 hours, respectively. What would be the chi-square statistic
represented by this test and the degrees of freedom of the statistic?
Solution: It is given that
χ2
=
(n − 1)=
S2 9 × 42
= 16
σ2 32
Since the chi-square statistic is based on S2 which has (10 −1) = 9 degrees of
freedom, therefore, it is distributed with (n −1) = 9 degrees of freedom.
x
1 − n
f (x) ( x )2
−1
n
e 2
; 0<x<∞
n
22
2 115
Block 1 Sampling Distributions
χ2 n
1
( ) ( )
− −1
f χ2
= n
e 2
χ2 2
; 0 < χ2 < ∞
n
2 2
2
n
where denotes the gamma function. The chi-square distribution has only
2
one parameter: a positive integer n that specifies the number of degrees of
freedom (It is also known as the shape parameter of the distribution because
the shape of the distribution depends on n as shown in Fig. 4.1). It means that
a chi-square distribution is determined by its degrees of freedom. There is a
different chi-square distribution for every value of df n. Therefore, it is a family
of continuous probability distributions.
Let us take an example to understand how to find degrees of freedom when
the pdf of a chi-square distribution is given.
Example 3: The probability density function (pdf) of a chi-square distribution is
given by
2
1 − χ2 2
( ) ( )
3
2
f χ
= e χ ; 0 < χ2 < ∞
96
What are the degrees of freedom of this chi-square distribution?
Solution: We know that the pdf of a chi-square distribution with n df is given
by
2
1 − χ2 2 n2 −1
= ( )
f χ2 n
n
e χ ; 0 < χ2 < ∞( )
22
2
We now compare the given pdf
2
1 − χ2 2
( ) ( )
3
f χ2
= e χ ; 0 < χ2 < ∞
96
with the standard form of the pdf of the chi-square distribution with n degrees
of freedom then we observe that
n
−1=3
2
Therefore,
n
= 4⇒n=8
2
After understanding the pdf of the chi-square distribution, we now study the
shape of the probability curve of the chi-square distribution and study the
116 effect of the degrees of freedom on the shape of it.
Unit 4 Sampling Distributions Associated with Normal Populations-I
Fig. 4.1: Chi-square probability density curves for (a) n = 1 and 2 (b) n = 3,5,10 and 20.
From Fig 4.1(b), you can see that as we increase the degrees of freedom the
curve tends to a normal curve.
4.3.2 Summary Measures of Chi-square Distribution
In the previous sub-section, the probability curve of the chi-square distribution
is discussed with some of its properties. Now, in this sub-section, we will
discuss some summary measures of the chi-square distribution.
The mean of the chi-square distribution with n degrees of freedom is its
degrees of freedom, that is, n. Hence,
Mean = n
Since the chi-square distribution is asymmetrical and right-skewed, therefore,
the mean is greater than the median and mode. The mode of the chi-square
distribution is given as
Mode = n – 2 when n > 2.
The variance of the chi-square distribution is 2n, that is,
Variance = 2n
Since the mean and variance of the chi-square distribution are n and 2n,
therefore, we can say that the variance of the distribution is double of its
mean.
117
Block 1 Sampling Distributions
The MSCAST programme is applied in nature, therefore, we do not give the
proof of the mean and variance of the chi-square distribution. If someone is
interested in that he/she can derive these summary measures as discussed in
MST-012.
Let us take some simple examples based on pdf and summary measures.
Example 4: For the probability density function (pdf) of the chi-square
distribution given in Example 3, find the mean and variance of this distribution.
Solution: In Example 3, we obtained the degrees of freedom of the given chi-
square distribution as n = 8. We know that the mean and variance of the chi-
square distribution with n degrees of freedom are
Mean = n and Variance = 2n
Therefore, for n = 8, we have
Mean = 8 and Variance = 16
Try the following Self Assessment Question to make sure that you have
understood the chi-square distribution.
SAQ 2
The pdf of a chi-square distribution is given as follows:
2
1 − χ2
=f χ
2
( )
e2
; 0 < χ2 < ∞
Obtain the degrees of freedom of the chi-square distribution. Also, find its
mean and variance.
1 − 2x 2
1 − x/2
f (x) ( )
−1
= = e x 2 e
2
n 2
2 2
118 2
Unit 4 Sampling Distributions Associated with Normal Populations-I
( )
If X ~ N μ,σ 2 , then Z =
X −μ
σ
~ N ( 0,1) and
2
X −μ
=Z2 2
~ χ(1)
σ
In general, the sum of squares of n standard normal variate follows a
chi-square distribution with n df, that is,
( )
If Xi ~ N μi ,σi2 , then Zi =
Xi − μi
σi
~ N ( 0,1)
Therefore,
n n 2
Xi − μi
= Zi2 ∑
∑ σi
2
~ χ(n)
=i 1 =i 1
SAQ 3
A stock market expert collected the share price of a Pharma company in 11
days which are given as follows:
70, 76, 85, 96, 102, 105, 100, 95, 88, 75, 72
It is assumed that the share price is normally distributed. The stock market
expert first standardises the price and then squares the prices. What will be
the distribution of the share price after the transformation? Also, find the mean
and variance of that distribution.
SAQ 4
List the applications of the chi-square distribution.
After understanding the chi-square distribution, we now come to the next
exact/standard sampling distribution, that is, the t-distribution.
4.7 t-DISTRIBUTION
The t-distribution is a continuous distribution which is very similar to the
standard normal distribution. The t-distribution was introduced by William
Sealy Gossett in 1908. Gossett was the Chief Brewer at the Guinness Brewery
in Dublin and was dedicated to applying the scientific method to beer
production. He needed a procedure for statistically analysing of small batches
of barley. In 1908, Gossett discovered the t-distribution for this purpose. The
(1876-1937) Guinness Brewery did not allow its workers to publish findings under their own
William Sealy Gossett was names so that the competitors would not learn about their methods. Therefore,
a scientist (chief Brewer) Gossett published his findings under the pen name “Student”. As a result,
at the Guinness brewery the distribution became known also as Student’s t-distribution.
in Dublin, Ireland.
As we discussed in Unit 2, when the population is normal and the population
120 standard deviation (σ) is not known, then we may use the sample standard
Unit 4 Sampling Distributions Associated with Normal Populations-I
deviation (S) in place of the population standard deviation (σ). But, due to the
discrepancy between sample SD (S) and population SD (σ) especially when
the sample standard deviation (S) is calculated from a very small sample, then
X−µ
the distribution of the statistic follows the t-distribution instead of the
S/ n
standard normal. Since it is associated with the sample standard deviation
which has (n − 1) degrees of freedom, therefore, the statistic
X−µ
t=
S/ n
follows the t-distribution with (n − 1) degrees of freedom.
Let us take an example to understand how to calculate the t-statistic and its
degrees of freedom.
Since the t-statistic is based on S2 which has (16 −1) = 15 degrees of freedom,
therefore, it is distributed with (16 −1) = 15 degrees of freedom.
After understanding the t-statistic, we now learn more about the t-distribution
such as its pdf and shape.
The t-distribution is a family of continuous probability distributions. Since it is
very similar to the standard normal distribution, therefore, the Student's
t-distribution plays a role in a number of widely used statistical analyses. We
will discuss its application in Section 4.10.
If X is a continuous random variable, then X follows a t-distribution with n
degrees of freedom if and only if it has the following probability density
function (pdf):
1
f (x) n +1
; −∞< x<∞
1 n x2 2
n B , 1 +
2 2 n
121
Block 1 Sampling Distributions
Symbolically, we denote that X follows a t-distribution with n degrees of
freedom as X ~ t (n) .
1
=f (t) n +1
; −∞<t<∞
1 n t 2 2
n B , 1 +
2 2 n
1 n
where B , is known as the beta function and
2 2
1 n
1 n 2 2 a b
=B , = B ( a,b )
2 2 n +1 a + b
2
n
where denotes the gamma function. As the chi-square distribution has
2
only one parameter (degrees of freedom), similarly, the t-distribution also has
only one parameter: a positive integer n that specifies the number of degrees
of freedom (It is also known as the shape parameter of the distribution
because the shape of the distribution depends on n as shown in Fig. 4.2).
Therefore, a t-distribution is determined by its degrees of freedom. For each
value of n, there is a different t-distribution. Therefore, it is a family of
continuous probability distributions.
Let us take an example.
Example 6: The probability density function (pdf) of a t-distribution is given
by
3
=f (t) 5/2
; −∞<t<∞
t2
8 1 +
4
We now compare the given pdf with the pdf of the t-distribution with n degrees
of freedom then we observe that
n +1 5
=
122 2 2
Unit 4 Sampling Distributions Associated with Normal Populations-I
Therefore,
n=4
After understanding the pdf of the t-distribution, let us see its shape and see
how the probability curve differs from the standard normal distribution and also
observe the effect of the degrees of freedom on its shape.
4.7.1 Probability Density Curve of t-distribution
The shape of the t-distribution depends on the degrees of freedom. In Fig. 4.2,
we have plotted the probability density curves of the t-distribution for n = 1, 5,
10, and 15 degrees of freedom along with the probability density curve of
standard normal distribution. By looking at the figure, you can observe that the
probability density curve of the t-distribution is bell-shaped and symmetric
about t = 0 line as the standard normal curve but it has a lower peak and
heavier tails (more observations near the tail, which means that it gives a
higher probability to the tails than the standard normal distribution) than the
standard normal curve. As the degrees of freedom increase, the curve pulls in
tighter around zero, the tails become thinner and the peak becomes taller. In
other words, we can also say that as the degrees of freedom increase, the t-
distribution will come closer to the standard normal distribution. Therefore,
the standard normal distribution can be used in place of the t-distribution with
large sample sizes.
Fig. 4.2: Probability curves for t-distribution for n =1, 5, 10 and 15 along with standard
normal curve.
But keeping the applied nature of the programme in view, we are not focusing
on proof of each measure. If someone is interested in that he/she can derive
these summary measures as discussed in MST-012.
After describing the t-distribution and its probability density curve and
summary measures, let us take an example.
Example 7: Consider the probability density function (pdf) of a t-distribution
given in Example 6 and find its mean and variance.
Solution: In Example 6, we obtained the degrees of freedom of the given t-
distribution as n = 4. We know that the mean and variance of the t-distribution
with n degrees of freedom as
n
Mean = 0 and Variance =
n−2
Therefore, for n = 4, we have
Mean = 0 and Variance = 2.
Now, try the following Self Assessment Questions for your practice.
SAQ 5
Obtain the degrees of freedom of a t-distribution whose pdf is given below:
1
=f (t) 3
; −∞<t<∞
1 5 t2
5 B , 1 +
2 2 5
Also, find the mean and variance of the above distribution.
As you have seen the probability density curve of the t-distribution is closely
related to the standard normal curve. Let us see the relationship of the t-
distribution with other well-known distributions.
SAQ 6
Show that the t-distribution with 1 degree of freedom is a standard Cauchy.
2. The t-distribution has only one parameter n, that is, the degrees of
freedom and the shape of the distribution depend on it. Therefore, there is
a different t-distribution for each value of df.
3. The mean of the t-distribution is zero, that is, mean = 0.
SAQ 7
Write any three applications of the t-distribution.
We now end this unit by giving a summary of what we have covered in it.
4.11 SUMMARY
A brief summary of what we have covered in this unit is given as follows:
• The degree of freedom is the total number of observations minus the
number of independent constraints or restrictions imposed on the
observations (or minus the number of estimated parameters).
• The probability density function of the chi-square distribution with n df is
given by
χ2 n
1
( ) (χ )
− −1
2 2 2
f χ
= n
e 2
; 0 < χ2 < ∞
n
2 2
2
• The mean and variance of the t-distribution with n degrees of freedom are
n
0 and , respectively.
n−2
58 + 75 + 51 + 69 + 52
X2 = 61
5
69 + 73 + 62 + 68
=X2 = 68
4
n
∑ ( X − X)
2
We now calculate SSD, that is, i
i =1
n
Calculation for ∑ ( Xi − X ) :
2
i=1
Deviation
SC Mean Deviation
Square
70 82 –12 144
94 82 12 144
67 82 –15 225
82 82 0 0
87 82 5 25
92 82 10 100
127
Block 1 Sampling Distributions
58 61 –3 9
75 61 14 196
51 61 –10 100
69 61 8 64
52 61 –9 81
69 68 1 1
73 68 5 25
62 68 –6 36
68 68 0 0
1069 1069 0 1150
1
5 +1
; −∞<t<∞
1 5 t 2 2
5 B , 1 +
2 2 5
After comparing the given pdf with the standard form, we get n = 5. So,
the degrees of freedom of the given t-distribution is 5.
6. We know that the probability density function of the t-distribution with n
df is given by
1
f (t) n +1
; −∞<t<∞
1 n t 2 2
nB , 1 +
2 2 n
11
1 1 1 2 2
=f (t) ; − ∞ < t < ∞ B , = = π π = π
π (1 + t 2 ) 2 2 1
which is the pdf of the standard Cauchy distribution. Hence, the Cauchy
distribution is a particular case of the t-distribution for n = 1.
7. Refer to Section 4.10.
χ2 n
1
( ) (χ )
− −1
2 2 2
f χ
= n
e 2
; 0 < χ2 < ∞
n
2 2
2
We have to find the pdf of the chi-square distribution when n = 5, so by
putting n = 5 in the above expression, we get 129
Block 1 Sampling Distributions
χ2 5
1
( ) (χ )
− −1
f χ2
= 5
e 2 2 2
; 0 < χ2 < ∞
5
2 2
2
χ2 3
1
(χ )
−
2 2
= 5
e 2
3 1 1
22 × ×
2 2 2
2
1 − χ2 2 3
1
= e χ ( ) 2
= π
π 2
3
2
2. We know that the probability density function of the t-distribution with n
df is given by
1
=f (t) n +1
; −∞<t<∞
1 n t2 2
nB , 1 +
2 2 n
1 3
+
2 2 ab
= = 2
B ( a,b )
1 3 t2 a+b
3 1 + 2
2 2 3
2 n n n
= 2 = − 1 − 1
1 1 1 t2 2 2 2
3 × 1 +
2 2 2 3
2 1
= 2
; −∞<t<∞ = π and 2 = 1
t2 2
3π 1 +
3
The value of the t-statistic can be calculated by the formula given below:
X −μ
t=
S/ n
Therefore, we have
290 − 300 −10
t= = = −1
50 / 25 10
130
UNIT 5
SAMPLING DISTRIBUTIONS
ASSOCIATED WITH
NORMAL POPULATIONS-II
Structure
5.1 Introduction 5.6 Tabulated Values of
t-distribution
Expected Learning Outcomes
5.7 Tabulated Values of Chi-
5.2 F-distribution
square Distribution
Probability Density Curve of
5.8 Tabulated Values of
F-distribution
F-distribution
Summary Measures of F-distribution
5.9 Summary
5.3 Relation of F-distribution to
5.10 Terminal Question
Other Distributions
5.11 Solutions /Answers
5.4 Properties of F-distribution
5.5 Applications of F-distribution
5.1 INTRODUCTION
In Unit 4, we have discussed the chi-square and t-distributions in detail with Tools You Will Need
their properties and applications. The F-distribution also has numerous real- The following terms are
world applications especially when discussing variance analysis and considered essential
hypothesis testing of the variances of the two normally distributed populations. background material for
For example, in finance, it is used to check whether the variances of stock this Unit. If you doubt your
returns are equal across two or more stocks. In engineering, it is also used to knowledge of any of these
test the effectiveness of different manufacturing processes by comparing the terms, you should review
variances of the outcomes. Additionally, the F-distribution is used in the appropriate Unit or
biostatistics to compare the variances of health outcomes across different section before proceeding:
treatments or interventions. In this unit, we discuss the F-distribution in detail • Sampling
and explain the method of reading the tabulated value of t, chi-square and F- Distributions for
distribution tables. Means and Variance
(Units 2 and 3).
This unit is divided into 11 sections. Section 5.1 is introductive. In Sections 5.2
to 5.5, we discuss the F-distribution with its probability density curve, summary
measures, relation to other distributions, properties and applications. As the
standard normal distribution has a standard normal table (Z-table), in a similar
131
Unit Write- Dr. Prabhat Kumar Sangal, School of Sciences, IGNOU, New Delhi
Block 1 Sampling Distributions
way, the chi-square, t and F-distributions also have their tables. Therefore,
Sections 5.6 to 5.8 are devoted to how to read tabulated values of the t, chi-
square, and F-distributions, respectively. The unit ends by providing a
summary of what we have discussed in this unit in Section 5.9. The terminal
questions and the solution of the SAQs/TQs are given in Sections 5.10 and
5.11, respectively.
In the next unit, we shall discuss the estimation of the unknown parameters.
5.2 F-DISTRIBUTION
The F-distribution is also a continuous probability distribution as the chi-square
or t-distribution. It is a sampling distribution that commonly occurs in statistics
when we discuss variances. As the chi-square distribution is not very useful in
describing real-world data, similarly, the F- distribution is not much used to
(1890 – 1962) describe real-world data. However, it is used for estimating and hypothesis
Sir Ronald Aylmer Fisher testing related to variances of two normal populations.
was a British and worked
as a statistician, The F-distribution was first introduced by British statistician Ronald A. Fisher in
mathematician, biologist, 1928. So sometimes it is called Fisher's F-distribution. Later on, George
geneticist, and academic. Waddel Snedecor was an American mathematician and statistician who
For his work in statistics, tabulated the F-distribution and used the letter F in Fisher's honour. The
he has been known as
distribution is also known as Snedecor's F distribution or the Fisher-
father of modern statistical
Snedecor distribution. Prof. Ronald. A. Fisher defined the F-distribution
sciences.
when he was interested in comparing the variances of two normally distributed
populations, and he derived the F-distribution as the ratio of two independent
chi-square variates when divided by their respective degrees of freedom, that
is,
χ(2n1 −1) / ( n1 − 1)
F=
χ(2n2 −1) / ( n2 − 1)
Thus, the sampling distribution of the ratio of sample variances follows the
F-distribution with (n1 –1, n2 – 1) degrees of freedom. The F-distribution is also
a family of distribution, and it has different shapes for each combination of
these degrees of freedom. Let us do an example to learn how to calculate the
F-statistic.
Example 1: Suppose a student of the MSCAST programme of Jammu Kasmir
region wants to test whether the variance of the weight of apples produced by
two different orchards in Kasmir is the same. He collects a random sample of
25 apples from Orchard I and 15 apples from Orchard II and obtains the
sample variances of 62 grams and 45 grams, respectively. Compute the value
of the F-statistic and also find the degrees of freedom associated with the F-
statistic.
Solution: Here, we are given that
=n1 25,
= n2 15,
S12 62,
= = S22 45
Since the F-statistic is associated with S12 and S22 , therefore, the degrees of
freedom of it depends on the degrees of freedoms of S12 and S22 . Since S12
has n1 – 1 = 25 – 1 = 24 degrees of freedom and, S22 has n2 – 1 = 15 – 1 = 14
degrees of freedom, therefore, the F-statistic has ( 24, 14) degrees of
freedom.
After understanding the F-statistic, let us discuss the probability density
function (pdf) of the F-distribution. As you seen, it has two degrees of freedom,
one for the numerator (n1) and the other for the denominator (n2), therefore,
the pdf of the F-distribution is more complex than the chi-square and the
t-distributions. We can define the F-distribution as
A continuous random variable X follows an F-distribution with (n1, n2 ) degrees
of freedom if and only if it has the following probability density function:
n1
n1 2 n1
−1
n
2 x 2
f ( x) ; 0<x<∞
n1 n2 n1 + n2
B , n1 2
2 2 1 + x
n2
By comparing the above form with the standard form, we get degrees of
freedom n1 = 2 and n2 = 2.
After understanding the form of the pdf of the F-distribution, you may be
interested to know the shape of the F-distribution and the impact of both
degrees of freedom on the shape of it. Let us discuss the probability density
134 curve of the F-distribution in the next section.
Unit 5 Sampling Distributions Associated with Normal Populations-II
Fig. 5.1: Probability density curves of F-distribution for various degrees of freedom.
In Fig. 5.1 (b), we plot different probability curves and put n1 as fixed at n1 = 5
and increase n2 = 5 to 10 to 20 to 50. By increasing the second parameter n2
from 5 to 50, the mean of the distribution (shown by the vertical line) 135
Block 1 Sampling Distributions
decreased, and the probability curve shifted from the tail to the centre of the
distribution. Similarly, in Fig. 5.1 (c), we plot different probability curves and
put n2 as fixed at n2 = 5 and increase n1 = 5 to 10 to 20 to 50. By increasing
the first parameter n1 from 5 to 50, the mean of the distribution (shown by the
vertical line) does not change but the probability curve shifted from the tail to
the centre of the distribution. After looking at the probability curves of the F-
distribution, we can observe that the probability curve of the F-distribution is a
uni-model curve for n1, n2 > 2.
After understanding the probability curve of the F-distribution with some of its
properties. In this sub-section, we will discuss some summary measures of it.
2n22 ( n1 + n2 − 2 )
Variance = for n2 > 4.
n1 ( n2 − 2 ) ( n2 − 4 )
2
Since both mean and variance of the F-distribution exist for n2 > 2 and n2 > 4,
respectively, therefore, for n1 = 2, and n2 = 2, the mean and variance of the F-
distribution do not exist.
Now, try to answer the following Self Assessment Question to see how much
136 you learn about the F-distribution.
Unit 5 Sampling Distributions Associated with Normal Populations-II
SAQ 1
If a random variable X follows the F-distribution whose pdf is given by
3 1
f ( x) 5
; 0<x<∞
8 x 1 2
1 + 4 x
then obtain the degrees of freedom of this distribution. Also, find its mean and
variance.
After understanding the F-distribution with its pdf, probability density curve and
summary measures, let us see the relationship of the F-distribution with other
well-known distributions.
SAQ 2
If a statistic t follows Student’s t-distribution with 4 df, then what will be the
distribution of the square of t? Also, write the pdf of that distribution.
SAQ 3
Write any five properties of the F-distribution.
SAQ 4
Write four applications of the F-distribution.
different degrees of freedom (n) such that the area under the probability curve t-value table, or t-test table.
of the t-distribution to its right tail (upper tail) is equal to α (α is also known as
level of significance in testing of hypothesis) as shown in Fig. 5.2 (a). In the t-
table, the first column of the left-hand side represents the degrees of freedom
(n) while the column heading represents the upper (right-hand side) tail
area/probability (α) of the probability curve of the t-distribution. The body
contains the value of the t-statistic for each particular value of n and α which
represents the critical value or point beyond which the area/probability of the t-
distribution is α. The area/probability (α) represents the proportions of the
t-distribution contained in the right tail.
To read the t-distribution table, you only need to know three values:
• the tail of the t-statistic on which α lies (one tail or two tails). 139
Block 1 Sampling Distributions
If the area/probability, that is, alpha (level of significance) of the t-statistic may
lie on the right tail, left tail or both tails so the tabulated value of the
t-distribution is called right tail value, left tail value or both tail values. We now
discuss how to read the tabulated values in each case one at a time.
then proceed right to the column headed α = 0.05. Find the cell in the table at
the intersection of degrees of freedom n = 6 and α =0.05 level. This is the
t-distribution value. For your convenience, we give a part of the t-table in Table
5.1.
Thus, we get the required right value of the t-statistic as t(n), α = t(6), 0.05 = 1.943.
The value of the t-statistic equal to 1.943 means, the probability that the
t-statistic would exceed (greater than) 1.943 is 0.05 as shown in Fig. 5.3.
Table 5.1: Part of t-table
Suppose in the above example, if we want to find the value of the t-statistic
such that the left area is 0.05 for 6 df then due to the symmetry of the t-
distribution the value of the t-statistic will be – t(n), α = – t(6), 0.05 = −1.943. The
tabulated value is shown in Fig. 5.4. Fig. 5.4
Both tails values
After learning how to read tabulated values for right and left tails, we now
discuss how to read it for two tails. For two-tails, there are two tabulated
values. To read both tails tabulated value, we follow the following steps:
Step 1: First of all, we half the total area on both tails. If it is α then the half
area i.e. α/2 lies in both tails as shown in Fig. 5.2(c).
Step 2: We then read the right-tail and left-tail tabulated values for the same 141
Block 1 Sampling Distributions
degrees of freedom (n) and α/2 (area/probability lie on right and left
tails) instead of α as discussed in the case of right and left tails
tabulated values.
We represent both tails tabulated values of the t-statistic for n degrees of
freedom and α level as ± t(n), α/2.
For example, if we want to find out the values of the t-statistic for which the
area on both tails is 0.05 and the degrees of freedom is 6. Since the total area
on both tails is 0.05, therefore, the area on the right tail as well as on the left
tail will be 0.05/2 = 0.025. Thus, we start with the first column of the t-table and
downward headed n until entry 6 is reached. Then proceed right to the column
headed α = 0.025. We get t(n), α/2 = t(6), 0.025 = 2.447. Since the t-distribution is
symmetrical at the t = 0 line, therefore, the left tail value will be the same as
the right tail value but in a negative sign. Therefore, – t(n), α/2 = – t(6), 0.025 = –
2.447. So the required values of the t-statistic are ± t(n), α/2 = ± t(6), 0.025 = ± 2.447.
Fig. 5.5 We also show the tabulated values in Fig. 5.5.
Example 4: Find the tabulated value of the t-statistic in each case for which
the degrees of freedom (n) and the area (level of significance) are given as
follows:
Solution:
(i) Here, we want to read the tabulated value of the t-statistic for
Therefore, we start from the first column of the t-table (Table V) given in
the Appendix at the end of this volume and downward headed n until
entry 10 is reached. Then proceed right to the column headed α = 0.01.
So we get the required tabulated value of the t-statistic as t(n), α = t(10), 0.01
= 2.764.
First of all, we read the right tail value by proceeding the same way as
part (i) t(n), α = t(8), 0.05 = 1.860. Since the t-distribution is symmetrical at the
t = 0 line, therefore, the left tail value is – t(n), α = – t(8), 0.05 = – 1.860.
(iii) Here, we want to read the value of the t-statistic for two tails and it is
given that
n = 14 and α = 0.10
Since the total area on both tails is 0.10, therefore, the area on the right
tail as well as on the left tail will be 0.05/2 = 0.05. Thus, we start with the
142 first column of the t-table and downward headed n until entry 14 is
Unit 5 Sampling Distributions Associated with Normal Populations-II
reached. Then proceed right to the column headed 0.025. We get t(n), α/2 =
t(14), 0.05 = 1.761. Since the t-distribution is symmetrical at the t = 0 line,
therefore, the left tail value will be the same as the right tail value but in a
negative sign. Therefore, – t(n), α/2 = – t(14), 0.05 =– 1.761. So the required
values of the t-statistic are ± t(n), α/2 = ± t(14), 0.025 = ± 1.761.
After understanding how to read the tabulated (critical) values for different
cases, we now discuss some more facts about the t-table. If we closely inspect
the t-table then we observe:
• The t-table contains only the right-tailed tabulated values.
• As the degrees of freedom increase, the tabulated value decreases. The
reason is that the tails of the t-distribution shift towards the centre as the
degrees of freedom increase as shown in Fig. 5.6.
• The t-distribution table does not include entries for every possible degree
of freedom. For example, the table lists continuously up to df = 30 and Fig. 5.6
after that t values for df = 40, 60, 120 and does not list entries for
degrees of freedom values between these.
Since the t-table does not include entries for every possible degree of
freedom, therefore, a question may arise “how do we read the tabulated
values for df which are not included in the t-table?” Don’t bother about
that we can read such values either using software or by the interpolating
method which is discussed as follows:
Method of Finding the Values of t-statistic for Degrees of Freedom which
are not Listed in the Table
The t-table (Table V) given in the Appendix does not list values for every
possible degree of freedom. Therefore, it becomes necessary to know how to
find values of the t-statistic for degrees of freedom not listed in the table. Let
us discuss the process of finding the values which are not listed in the table
with the help of an example.
Suppose we want to find out the tabulated value of the t-statistic for 34
degrees of freedom which is not listed in the table such that the area on the
right side is equal to 0.05.
For that, first of all, we read the tabulated values of the t-statistic that are just
greater and just smaller than the degrees of freedom for our interest. Thus,
from the t-table, we get the values of t-statistic for 40 and 30 degrees of
freedom and area α = 0.05 as
t(40), 0.05 = 1.684 and t(30), 0.05 = 1.697
Note that the larger the degrees of freedom, the smaller the tabulated value of
the t-statistic.
We now calculate how much the t-value changes for each degree of freedom
between these two tabulated values. Here, there is a difference of 10(40 – 30)
degrees of freedom and a t-value change of 0.013(1.697 – 1.684).
Thus, we can obtain the change in the t-value corresponding to a unit change
in degree of freedom as 143
Block 1 Sampling Distributions
0.013
= 0.0013
10
Since we have to obtain the value for 34 degrees of freedom, this is either 4
more than 30 or 6 less than 40. Therefore, we can interpolate from either
value. To get from 30 to 34 degrees of freedom there is a difference of 4 (34 –
30). So we multiply this difference by the amount by which the t-value changes
per degree of freedom i.e. 0.0013. This result as
4 × 0.0013 =
0.0052
Since the larger the degrees of freedom the smaller the tabulated value of the
t-statistic, therefore, subtracting this value 0.0052 from t(30), 0.05 = 1.697 to get
the required value. Thus,
SAQ 5
Find the tabulated values of the t-statistic for which the area and the degrees
of freedom are given as follows:
(i) n = 9 and α = 0.01 (right tail)
In the chi-square table, the column headings indicate the area on the upper
portion (right tail) of the probability curve of the chi-square distribution and the
first column on the left-hand side indicates the values of degrees of freedom
(n). The body contains the value of the chi-square statistic for each particular
value of n and α which represents the critical value or point beyond which the
area/probability of the chi-square distribution is α. The area/probability (α)
represents the proportions of the chi-square distribution contained in
144 the right tail.
Unit 5 Sampling Distributions Associated with Normal Populations-II
The procedure to read the right tail tabulated value of the chi-square statistic is
almost similar to reading the t-table, we follow the following steps:
Step 1: We start with the first column of the chi-square table, that is degrees
of freedom and downward headed ‘n’ until the required degree of
freedom is reached.
Step 2: After that, we proceed right to the column headed α up to the
required α is reached.
Step 3: We get the required right tail tabulated value in the cell of the table at
the intersection of required degrees of freedom n and α.
We represent the tabulated value of the chi-square statistic for n degrees of
2
freedom and for right tail area/probability as χ (n), α .
Suppose we want to find out the tabulated value of the chi-square statistic for
which the area/probability on the right tail is 0.01 and the degrees of freedom 145
Block 1 Sampling Distributions
is 4. To read the tabulated value, we start with the first column of the chi-
square table, that is degrees of freedom and downward headed ‘n’ until entry 4
is reached and then proceed right to the column headed α = 0.01. Then we
find the cell in the table at the intersection of degrees of freedom n = 4 and α =
0.01 level. For your convenience, we give a part of the chi-square table as
shown in Table 5.2.
Table 5.2: The Part of Chi-square Table
α= 0.995 0.99 0.975 0.95 0.90 0.10 0.05 0.025 0.01 0.005
n(df)=1 --- --- --- --- 0.02 2.71 3.84 5.02 6.63 7.88
2 0.01 0.02 0.05 0.10 0.21 4.61 5.99 7.38 9.21 10.60
3 0.07 0.11 0.22 0.35 0.58 6.25 7.81 9.35 11.34 12.84
4 0.21 0.30 0.48 0.71 1.06 7.78 9.49 11.14 13.28 14.86
5 0.41 0.55 0.83 1.15 1.61 9.24 11.07 12.83 15.09 16.75
6 0.68 0.87 1.24 1.64 2.20 10.64 12.59 14.45 16.81 18.55
Suppose we want to read the tabulated value for which the left area is 0.05
and the degree of freedom is 4. To read this tabulated value, we start with the
first column of the chi-square table, that is degrees of freedom and downward
headed ‘n’ until entry 4 is reached and then proceed right to the column
headed α to 1 – 0.05 = 0.95 instead of 0.05. Then we find the cell in the table
146 at the intersection of degrees of freedom n = 4 and α = 0.95 level. Thus, we
Unit 5 Sampling Distributions Associated with Normal Populations-II
2
get the required left value of the chi-square statistic as χ (n),(1−α ) = χ2(4),0.95
= 0.71. The tabulated value of the chi-square statistic equal to 0.71 means,
the probability that the chi-square statistic would be less than 0.71 is 0.05 and
greater than 0.71 is 0.95 as shown in Fig. 5.9.
Let us learn how to read both/two tails tabulated values from the chi-square
table.
Both tails values
The procedure to read the two tails tabulated value of the chi-square statistic Fig. 5.9
is almost similar to the t-table, we follow the following steps:
Step 1: First of all, we half the total area on both tails. If it is α then the half
area i.e. α/2 lies in both tails as shown in Fig. 5.7(c).
Step 2: We then read the right tail and left tail tabulated values for the same
degrees of freedom (n) and α/2 (area/probability lie on right and left
tails) instead of α as discussed in the case of right and left tails
tabulated values.
2
In this case, we represent the right-tail tabulated value as χ (n), α /2 and left-tail
2
as χ (n),(1−α /2) .
For example, if we want to read the values of the chi-square statistic for which
the area on both tails is 0.05 and the degrees of freedom is 6. For that first of
all, we half the total area/probability as 0.05/2 = 0.025. After that, we read the
right tail and left tail values in such a way that the area on both the tails
remains α/2 = 0.025 as discussed in the case of right and left tails tabulated
2
values cases. We get χ (n), α /2 = χ2(6),0.025 =
14.45 and χ2(n),(1−α /2) = χ2(6),(1−0.025)
χ2(6),0.975 =
= 1.24 . These values are shown in Fig. 5.10.
Fig. 5.10
After understanding how to read tabulated (critical) values for different cases,
we now discuss some more facts about the chi-square table. If we closely
inspect the chi-square table, then we observe:
• The table contains the tabulated values for the right tail as well as the left
tail.
• As the degrees of freedom increase, the tabulated value decreases. The
reason is that as we increase the degrees of freedom the tails of the chi-
square distribution shift towards the right side as shown in Fig. 5.11.
• The chi-square table does not include entries for every possible degree of
freedom. For example, the table lists continuously up to df = 30 and after
that the chi-square values for df = 40, 60, 120 and does not list entries for
degrees of freedom values between these. Fig. 5.11
Example 5: Find the values of the chi-square statistic for which the degrees
of freedom (n) and area are given as follows:
(i) n = 2 and α = 0.05 (right tail)
(ii) n = 10 and α = 0.01 (left tail)
(iii) n = 8 and α = 0.05 (both tails)
(iv) n = 64 and α = 0.01 (right tail)
Solution:
(i) Here, we want to read the tabulated value of the chi-square statistic for
n = 2 and α = 0.05 (right tail)
Thus, we start from the first column of the chi-square table given in the
Appendix and downward headed n until entry 2 is reached. Then
proceed right to the column headed α = 0.05. We get the required
2 2
tabulated value of the chi-square statistic as χ(n),α =
χ(2),0.05 5.99.
=
(ii) Here, we want to find the value of the chi-square statistic for
n = 10 and α = 0.01 (left tail)
To read the left-tail tabulated value, we start with the first column of the
chi-square table, that is, degrees of freedom and downward headed ‘n’
until entry 10 is reached and then proceed right to the column headed α
at 1 – 0.01 = 0.99 instead of 0.01. Then we find the cell in the table at the
intersection of degrees of freedom n = 10 and α = 0.99 level. Thus, we
get the required left value of the chi-square statistic as χ2(n),(1−α ) = χ2(10),0.99
= 2.56 .
(iii) Here, we want to find the tabulated value of the chi-square statistic for
two-tail area α = 0.05 and n = 8.
Since the total area on both tails is 0.05, therefore, half area 0.05/2 =
0.025 lies on both tails. In this case, the chi-square statistic has two
values one on the right-tail as χ(2n), α /2 =
χ(28 ),0.025 and one on the left-tail as
χ(2n),(1−α /2) =
χ(28 ),0.975 . So by proceeding the same way as above, we get the
required values of the chi-square statistic as χ(28 ),0.025 =
17.53 and
χ(28 ),0.975 =
2.18 .
(iv) Here, we want to find the tabulated value of the chi-square statistic for
the right tail and for n = 64 and α = 0.01.
Since the chi-square table does not have the tabulated value for 64
degrees of freedom so we need to interpolate it. For this, we find the
tabulated values of the chi-square statistic that are just greater and just
less than the degree of freedom 64 with α = 0.01. Thus, we have
Since the larger the degrees of freedom larger the tabulated value of the
chi-square statistic, so adding 4.816 in 88.38, we get the required value
as
χ(264 ),0.01 = 88.38 + 4.816 = 93.196
SAQ 6
Find the tabulated value of the chi-square statistic for which the degrees of
freedom and area are given as
(i) n = 11 and α = 0.01 (right tail)
(ii) n = 19 and α = 0.10 (left tail)
different tabulated value, therefore, the statisticians organised the F-table also known as the F-table,
somewhat differently than the tables for the other distributions. The F-score table, F-value table,
statisticians prepare a separate F-table that is associated with different α (the or F-test table.
area in the right tail of the distribution) and in each of these tables, the F
values are given for various combinations of degrees of freedom. Table VII in
the Appendix contains the tabulated (critical) values of the F-statistic for
various degrees of freedom such that the area under the probability curve of
the F-distribution to its right (upper) tail is equal to α = 0.10, 0.05, 0.025, 0.01
and 0.005. A part of the table is shown in Table 5.3.
In the F-table, the first row of the F-table indicates the values of degrees of
freedom for the numerator (n1) and the first column on the left-hand side
indicates the values of degrees of freedom for the denominator (n2). The body
of the table contains the value of the F-statistic for each particular pair of
(n1, n2) and α which represents the critical value or point beyond which the
area/probability of the F-distribution is α. The area/probability (α) represents
the proportions of the F-distribution contained in the right tail.
Suppose we want to find out the tabulated value of the F-statistic for which the
area/probability on the right tail is 0.05 and the degrees of freedom for the
numerator is 8 and for the denominator is 12. To read the tabulated value, we
first select the F-table corresponding to α = 0.05. Then we start with the first
Fig. 5.13
row of the selected F-table, that is degrees of freedom for the numerator and
proceed right up to required n1 = 8 is reached and then downward headed
150 ‘denominator degrees of freedom (n2)’ until required n2 = 12 is reached. The
Unit 5 Sampling Distributions Associated with Normal Populations-II
F(n1,n2 ), α F=
= (8,12),0.05 2.85 and is
shown in Fig. 5.13. For your convenience, we give a part of the F-table as
follows:
Table 5.3: Part of F-table for α = 0.05
Degrees of
freedom for Degrees of freedom for numerator (n1)
denominator
(n2) 1 2 3 4 5 6 7 8 9 10 11 12
10 4.96 4.10 3.71 3.48 3.33 3.22 3.14 3.07 3.02 2.98 2.94 2.91
11 4.84 3.98 3.59 3.36 3.20 3.09 3.01 2.95 2.90 2.85 2.82 2.79
12 4.75 3.89 3.49 3.26 3.11 3.00 2.91 2.85 2.80 2.75 2.72 2.69
13 4.67 3.81 3.41 3.18 3.03 2.92 2.83 2.77 2.71 2.67 2.63 2.60
14 4.60 3.74 3.34 3.11 2.96 2.85 2.76 2.70 2.65 2.60 2.57 2.53
15 4.54 3.68 3.29 3.06 2.90 2.79 2.71 2.64 2.59 2.54 2.51 2.48
16 4.49 3.63 3.24 3.01 2.85 2.74 2.66 2.59 2.54 2.49 2.46 2.42
Note 1: The value of the F-statistic (n1, n2) is not the same for (n2, n1) for the
denominator, that is,
F(n1,n2 ) ≠ F(n2 ,n1 )
It means that we can find the left-tail tabulated value of the F-statistic using the
right-tail values by interchanging the degrees of freedom. We represent the left
tail tabulated/critical value of the F-table as F(n1,n2 ),(1−α ) . To read the left tail
tabulated value of the F-statistic, we follow the following steps:
Step 1: First of all, we select the table of required α because there is a
separate F-table for each α.
Step 2: After selecting the required F-table, we first read the tabulated value
for the reverse degrees of freedom (n2, n1), that is, F(n2 ,n1 ), α as
discussed above.
Step 3: After that, we use the relationship of the F-distribution to find the left-
tail tabulated values
1
F(n1,n2 ),(1−α ) =
F(n2 ,n1 ), α
Suppose we want to find out the tabulated value of the F-statistic for which the
area/probability on the left tail is 0.01 and the degrees of freedom for the
numerator is (n1) 6 and for the denominator (n2) 10. To read the tabulated
value, we first select the F-table corresponding α = 0.01. After that we first 151
Block 1 Sampling Distributions
read the tabulated value for reversing the degrees of freedom, that is,
F(n2 ,n1 ), α = F(10,6),0.01 as discussed above. We get F(10,6),0.01 = 7.87 . After that, we
use the relationship to calculate the required left-tail value as
1
F(n1,n2 ),(1− α) =
F(n2 ,n1 ),α
1 1
F( 6,10 ),(1−=
0.01)
F( 6,10 ),=
( 0.99 ) = = 0.127
F(10,6 ),0.01 7.87
For example, if we want to read the values of the F-statistic for which the area
on both tails is 0.05 and the degrees of freedom (4, 12). For that first of all, we
half the total area/probability as 0.05/2 = 0.025. After that, we read the right-tail
and left-tail values in such a way that the area on both the tails the remaining
α/2 = 0.025 as discussed in the case of right and left tails tabulated values then,
1 1
we get F =(n1,n2 ), α /2 F=
(4,12),0.025 4.12 and F(n=
1 ,n2 ),(1−α )
F=
( 4,12),0.975 =
F(12,4),0.025 8.75
= 0.11.
152 • The F-table does not include entries for every pair of possible degrees of
Unit 5 Sampling Distributions Associated with Normal Populations-II
(ii) Here, we want to find the tabulated value of the F-statistic for which
n1 = 7, n2 = 4 and area/probability on the left tail is α = 0.05.
Since we have to read the tabulated value for the left tail, so we use the
relation:
1
F(n1,n2 ),(1−α ) =
F(n2 ,n1 ), α
We first select the F-table corresponding α = 0.05. After that we read the
tabulated value by reversing the degrees of freedom, that is, for (4, 7).
We get F =(n2 ,n1 ), α F=
(4,7),0.05 4.12. After that, we use the relationship to
calculate the required left-tail value as
1
F(n1,n2 ),(1−α ) =
F(n2 ,n1 ), α
1 1
F( 7,4 ),(1−=
0.05 )
F( 7,4=
),0.95 = = 0.243
F( 4,7 ),0.05 4.12
(iii) Here, the total area on both tails is 0.10, therefore, half area 0.10/2 =
0.05 lies on both tails. In this case, the F-statistic has two values one on
right-tail as F(n1,n2 ), α /2 = F(5,10),0.05 F(n1,n2 ),(1−α /2)
153
Block 1 Sampling Distributions
= F=
( 5,10 ),(1−0.05 ) F( 5,10 ),0.95 . So, by proceeding the same way as above, we
get the required tabulated values of the F-statistic as F( 5,10 ),0.05 = 3.33 and
1 1
F( 5,10=
),0.95 = = 0.21 .
F(10,5 ),0.05 4.74
(iv) Here, we want to find the tabulated value of the F-statistic for
n1 = 10, n2 = 32 and α = 0.01(right tail)
Since the F-table for α = 0.01 does not have the tabulated value
corresponding to degrees of freedom n1 = 10, n2 = 32 so we need to
interpolate it.
For this, we find the tabulated values of the F-statistic that are just
greater and just less than the degree of freedom 32. Thus, we have
F(10, 40) , 0.01 = 2.80 F(10, 30) , 0.01 = 2.98
There is a difference of 10 degrees of freedom between these two and a
difference of 0.18 (2.98 – 2.80) in the F-values. Thus, each degree of
freedom has an approximate change in the value of the F-statistic as
0.18
= 0.018
10
To get 32 degrees of freedom, multiplying 0.018 by 2 (32 – 30), we get
0.018 × 2 =0.036
Since the larger the degrees of freedom smaller the tabulated value of
the F-statistic, therefore, subtracting this from 2.98, we get the required
values as
F(10, 32) = 2.98 – 0.036 = 2.944
Now, try to answer the following Self Assessment Question.
SAQ 7
Find the tabulated value of the F-statistic in the following cases:
(i) n1 = 8, n2 = 10 and α = 0.05 (right tail)
(ii) n1 = 4, n2 = 12 and α = 0.01 (left tail)
Let us end with a brief look at what we have covered in this unit.
5.9 SUMMARY
In this unit, we have covered the following points:
• The probability density function of the F-distribution with (n1, n2 ) degrees
of freedom is given by
n1
n1 2 n1
−1
n 2 F 2
=f (F) ; 0<F<∞
n1 n2 n1 + n2
B , n1 2
2 2 1 + F
154 n2
Unit 5 Sampling Distributions Associated with Normal Populations-II
• The mean and variance of the F-distribution with (n1, n2) df are given as
follows:
n2
Mean = for n2 > 2.
n2 − 2
2n22 ( n1 + n2 − 2 )
Variance = for n2 > 4.
n1 ( n2 − 2 ) ( n2 − 4 )
2
And show that the pdf of the F-distribution is not the same by reversing the
degrees of freedom.
We now try to convert the given pdf in the form of the standard form of
the F-distribution so that we can compare and find the degrees of
5
x 2 1
freedom. Since the coefficient of x in 1 +
4 4
155
Block 1 Sampling Distributions
n n
that n1 = 1 and n2 = 4 because the coefficient of x in 1 + 1 x is 1 . We
n2 n2
try to write the given pdf into the F-distribution as follows:
1
1 2 1
4 x2
−1
=f ( x) ; 0<x<∞
1 4 1+ 4
B , 1 2
2 2 1 + 4 x
1 1
1 2 2
B =
,2 2
= 2 = 4
2 5 3 1 1 3
× ×
2 2 2 2
By comparing the above form with the standard form, we get n1 = 1 and
n2 = 4.
We know that the mean and variance of the F-distribution with (n1, n2)
degrees of freedom are:
n2 2n22 ( n1 + n2 − 2 )
Mean = for n2 > 2 and Variance = for n2 > 4.
n2 − 2 n1 ( n2 − 2 ) ( n2 − 4 )
2
Since the variance of the F-distribution exists for n2 > 4, therefore, for
n1 = 4 it does not exist.
2. As we know if a variable t follows the t-distribution with n df, then the
square of t follows the F-distribution with (1, n) df, therefore, for n = 4,
the distribution of t square follows the F-distribution with (1, 4) df. The
pdf of the F-distribution is given by
n1
n1 2 n1
−1
n2 F2
=f (F) ; 0<F<∞
n1 n2 n1 + n2
B , n1 2
2 2 1 + F
n2
1
1 2 1 1
4 F2
−1 −
3F 2
1 4
=f (F) = = Since B 2 ,2 3
1 4 1+ 4 5
B , 1 2 1 2
2 2 1 + F 8 1 + F
4 4
(ii) Here, we want to find the value of the chi-square statistic for
n = 19 and α = 0.10 (left tail)
To read the left-tail tabulated value, we start with the first column
of the chi-square table, that is degrees of freedom and downward
headed ‘n’ until entry 19 is reached and then proceed right to the
column headed α at 1 – 0.1 0= 0.90 instead of 0.10. Then we find
the cell in the table at the intersection of degrees of freedom n = 19
and α = 0.90 level. Thus, we get the required left value of the chi-
2
χ2(19),0.90 =
square statistic as χ (n),(1−α ) = 11.65.
Thus, we first select the F-table for α = 0.05 and then we start with
the first row of the selected F-table, that is, degrees of freedom for
157
Block 1 Sampling Distributions
(ii) Here, we want to find the tabulated value of the F-statistic for which
n1 = 4, n2 = 12 and area/probability on the left tail is α = 0.01.
Since we have to read the tabulated value for the left tail, so we use
the relation
1
F(n1,n2 ),(1−α ) =
F(n2 ,n1 ), α
1 1
F( 4,12),=
(1−0.01) F( 4,12=
),( 0.99 ) = = 0.07
F(12,4 ),0.01 14.37