Normal Distribution
Normal Distribution
A normal distribution is one in which the majority of cases are located in the middle
of the scale and only a small number of cases are located at both extremes of the
scale. A normal curve is one, which graphically represents normal distribution. Most
distributed. In reality, however, no distribution ever takes the absolute form of normal
distribution. Many frequency distributions are close to this absolute form and we
A normal probability curve is a bell shaped curve. The major characteristics of a normal
In a normal probability curve the X axis represents the trait values as “z”
proportion etc.
A normal curve is always symmetrical, that is, the right of the curve is
In a normal curve the mean, the median and the mode are numerically
A normal curve never touches the base line, no matter how far the curve is
stretched.
In a normal curve the highest ordinate is at the center. All ordinates on both
104
The scores of the trait under measurement are converted to “z” scores. It is
The range of the normal probability curve (z scores) is from –3σ to +3σ with
standard deviation unit taken on each side of the mean includes a total area of
34.13% + 34.13% = 68.26% of the curve. The range between –2σ to +2 σ includes
the area 47.72% + 47.72%= 95.44%. The range between +3σ to –3σ includes the
area 49.865% + 49.865% = 99.73%. this shows that where as the area between 0 to
1σ is 34.13%, the area between 1σ and 2σ is 13.59% and the area between 2σ and
3σ is 2.15% indicating that the maximum concentration of scores are there near the
scores reduce. While referring to the normal probability tables it should be noted
that all the areas given are for the z scores from mean that is 0.
Not all distributions are normally or nearly normally distributed. The deviations
105
to the normal distribution may occur in two ways. (1) skewness (2) Kurtosis.
8.4.1. Skewness : The distribution is said to be “skewed” when the mean and the
median fall at different point in the distribution, and the balance (or center of gravity)
is shifted to one side or the other (to left or right). In a normal distribution, the mean
equals the median exactly and the skewness is, of course, zero.
Distributions are said to be skewed negatively or to the left when scores are
massed at the high end of the scale (the right end) and are spread out more
Median
Mean an
Figure 8.2: Negative skewness: to the left
Distributions are skewed positively or to the right when scores are massed at
the low (or left) end of the scale, and are spread out gradually toward the high or
Median Mean
Figure.8.3. Postitive skewness: to the right
The greater the gap between mean and median, the greater the skewness.
106
When skewness is negative, the mean lies to the left of the median i.e. mean value
is lower than the median value; and when skewness is positive, the mean lies to the
right of the median i.e. the mean value higher than the median value.
8.4.2. Kurtosis
distribution as compared with the normal. A frequency distribution more peaked than
the normal is said to be leptokurtic; one flatter than the normal, platykurtic. Figure-
8.4. shows a leptokurtic distribution and a platykurtic distribution plotted on the same
diagram around the same mean. A normal curve (called mesokurtic) also has been
drawn in on the diagram to bring out the contrast in the figure, and to make
comparison easier.
Figure.8.4. Leptokurtic (A), Normal or mesokurtic (B) and Platykurtic (C) curves
generalization about the population from which the sample was drawn.
107
The primary objective of statistical inference is to enable us to generalize from
a sample to some larger population of which the sample is a part. A good sample is
measures obtained from the sample may not be identical to the population
measures. The sample measures may be slightly larger or smaller than the
population values. The degree to which the sample measures represent the
computed sample measures. Any deviation of such sample measure from the
population may be regarded as an error of estimation the standard error tells us how
large these errors of estimations are in any particular sampling situations. Therefore
sample measures. To distinguish such a standard deviation from the more familiar
samples from the population and compute means for all these samples the standard
deviation of these means is the standard error of mean for the samples drawn from
the population.
population a standard error is computed from one sample drawn from the
population.
that statistic. In this chapter we shall see the computation of standard error for
108
8.6.1. Standard error of mean
The formula for the standard error of the mean is
σ
SEM or σM =
N
σ = the standard deviation of the population
N = the number of cases in the sample.
1.858Q
SE Mdn =
N
.71σ
SEσ =
N
reasoning for 225 boys in the tenth class of a city was 27.26 with a SD of 11.20.
How dependable is this mean? Specifically, how good an estimate is it of the mean
which could be expected if all of the tenth class boys in city were tested.
109
11.20
SEM = = .75
225
of number of sample means around the fixed population mean. The normal curve in
the figure shows this sampling distribution. At its center (unknown) is the population
mean and its SD is 0.75. The sample mean may fall equally often on the positive
and negative sides of the population mean. A 0.05 level of significance means to
obtain a range where the sample mean is likely to be 47.5% on the positive side of
the population mean and 47.5% on the negative side of the population mean and the
chance factor is likely to occur 2.5% on positive side and 2.5% on the negative side.
From the normal probability tables 47.5% corresponds to +1.96 σ and -1.96 σ. It
means that 95 times out of 100 times the sample mean will not miss the population
mean by + 1.96 X 0.75 i.e. 1.47. Therefore the 95% confidence interval for the
population mean is 27.26 + 1.47, which is 25.79 and 28.83. So it means that if all
the tenth class students of the city were tested we can be confident 95 out of 100
times that it will be between 25.79 and 28.83. Similarly 99% confidence interval can
be obtained by looking into the normal probability tables for + 49.5 which is + 2.58.
8.7. Significance
upon the probability that the given difference could have arisen “by chance.”
110
Experimenters and research workers have for convenience chosen several
arbitrary standards called levels of significance of which the .05 and .01 levels are
most often used. A 0.05 level of significance means that the chance factor exists
2.5% on the positive side and 2.5% on the negative side of the normal probability
curve and 95% of the time the statistical value obtained is true. A 0.01 level of
significance means that the chance factor exists 0.5% on the positive side and 0.5%
on the negative side of the normal probability curve and 99% of the time the
statistical value obtained is true. The confidence with which an experimenter rejects
Experimenters have found the null hypothesis is a useful tool in testing significance.
In its simplest form this hypothesis asserts that there is no true difference between
two population means and that any difference found between sample means is
therefore accidental and unimportant. The null hypothesis is similar to legal principle
A one tailed test is a directional test, which indicates the direction of the
difference between the samples under study. Here the change can occur only in one
onwards the change is always in the positive direction which is the language
develops with the progress of each year. Here we are concerned whether the
change is significant or not. A two-tail test is one where the rejection of the null
hypothesis can be in either direction. While comparing two teaching methods A and
B if it is found that there is no difference the null hypothesis is accepted. But if the
111
null hypothesis is to be rejected it may be found that method A is better than B or
vice versa.
difference not significant, when a true difference actually exists. If we set up a high
least .01 in most experimental research, i.e., to risk type II errors by preventing those
of type I. But the .05 level is often satisfactory, especially in preliminary work.
8.12. Source
1. Garrett, H.E. (1966) Statistics in Psychology and Education. Bombay : Vakils
Feefer & Simon Pvt. Ltd.
2. Guilford J.P. (1965). Fundamental Statistics in Psychology and Education
(4th Edn.). New Delhi: Subject Publications.
3. Mc.Guigan, F.J. (1990) Experimental Psychology New Delhi: Prentice Hall of
India Limited.
4. Singh A.K. (1997) Tests measurements and Research Methods in
Behavioural science Patna: Bharati Bhavan Publishers and Distributors.
112