Statistics For Economists ch1 - ch7 (All in One)
Statistics For Economists ch1 - ch7 (All in One)
Statistics For Economists ch1 - ch7 (All in One)
ECON 2042
Addis Ababa University
Department of Economics
May, 2023
11/10/2023 Statistics for Economics_AAU 1
CHAPTER ONE
OVERVIEW OF BASIC PROBABILITY
THEORY
Introduction
❖ There are large numbers of happenings in nature and in the
realm of human activity that are associated with uncertainties.
❖ Uncertainty plays an important role in our daily lives and
activities as well as in business.
❖ It is often necessary to "guess" about the outcome of an event
in order to make a decision.
❖ In our daily lives we are faced with a lot of decision-making
situations that involve uncertainty.
11/10/2023 Statistics for Economics_AAU 2
For example:
✓ In a business context: investment, stock prices
✓ Game Wins, Lottery wins, Election win, Traffic jam
✓ Market Conditions: effectiveness of an ad campaign or the
eventual success of a new product.
✓ Weather forecast: appearance of clouds in the sky next morning
is not as certain.
✓ The sex of a baby to be born some months hence is again not
known for certain.
Types of Events
✓ We defined have already defined Event as any subset of
outcomes of an experiment.
1) Mutually Exclusive Events: If two or more events cannot occur
simultaneously in a single trial of an experiment, then such events
are called mutually exclusive events or disjoint events.
11/10/2023 Statistics for Economics_AAU 9
✓ In other words, two events are mutually exclusive if the occurrence
of one of them prevents or rules out the occurrence of the other.
For example, the numbers 2 and 3 cannot occur simultaneously on the
roll of a dice.
✓ Symbolically, a set of events {𝑨𝟏 , 𝑨𝟐 , . . ., 𝑨𝒏 } is mutually
exclusive if 𝑨𝒊 ∩ 𝑨𝒋 = ∅ for i ≠ j.
✓ This means the intersection of two events is a null set (∅); it is
impossible to observe an event that is common in both 𝑨𝒊 and 𝑨𝒋 .
C) Subjective Definition
✓ If there is little or no past experience on which to base probability it may be
guessed subjectively.
✓ Essentially, this means evaluating the available information and then estimating
the probability of an event.
✓ The subjective approach is always based on the degree of beliefs, convictions, and
experience concerning the likelihood of occurrence of a random event.
Example: Estimating:
i) The probability that a new product of a firm will be successful in the market.
ii) The probability that a student will score an A in the course, and
iii) When a person says that the probability of rain tomorrow is, say 70% - he is
expressing his personal degree of belief.
11/10/2023 Statistics for Economics_AAU 27
1.3. Axioms of Probability: The Rule of probability
✓ Before giving axiomatic definition of probability let’s first see, the summary of what
we know about set theory which are relevant to our interest.
De Morgan’s Law
ഥ∩𝑩
I) (𝐀 ∪ 𝑩)= 𝑨 ഥ
NOTE: Answers for questions iii to v provide us a proof for the third
axiom.
Exercises: The events X and Y are mutually exclusive. Suppose P(X) = .05
and P(Y) = .02.
i) What is the probability of either X or Y occurring?
ii) What is the probability that neither X nor Y will happen?
11/10/2023 Statistics for Economics_AAU 37
Rules of Multiplication
5) The Special Rule of Multiplication
✓ The special rule of multiplication requires that two events A and B are
independent.
✓ Two events are independent if the occurrence of one event does not alter
the probability of the occurrence of the other event.
Example: suppose two coins are tossed. The outcome of a coin toss (head or
tail) is unaffected by the outcome of any other prior coin toss (head or tail).
❑ For two independent events A and B, the probability that A and B will both
occur is found by multiplying the two probabilities.
❑ This is the special rule of multiplication and is written symbolically as:
𝑷(𝑨 𝒂𝒏𝒅 𝑩) = 𝑷(𝑨)𝑷(𝑩)
❑ For three independent events, A, B, and C, the special rule of multiplication
used to determine the probability that all three events will occur is:
𝑷(𝑨 𝒂𝒏𝒅 𝑩 𝒂𝒏𝒅 𝑪) = 𝑷(𝑨)𝑷(𝑩)𝑷(𝑪)
3) Combination Rule
✓ Sometimes the ordering or arrangement of objects is not important, but
only the objects that are chosen. For example, we may not care in what
order the books are placed on the shelf, but only which books you are able
to shelve. In addition, when a five-person committee is chosen from a
group of 10 students, the order of choice is not important because all 5
students will be equal members of committee.
❑ This counting rule for combinations allows us to select r (say) number of
outcomes from a collection of n distinct outcomes without caring in what
order they are arranged. This rule is denoted by
𝑛 𝒏!
nCr 𝒐𝒓 𝑟 =
11/10/2023 (𝒏 − 𝒓)! 𝒓!
Statistics for Economics_AAU 47
Example 1: From eight persons, how many committees of 3 members
may be chosen?
Solution: Since two committees are the same if they are made up of the same
members we have,
8 𝟖! 𝟖!
C
8 3 𝒐𝒓 = = = 𝟓𝟔 𝑑𝑖𝑓𝑓𝑒𝑟𝑒𝑛𝑡 𝑐𝑜𝑚𝑚𝑖𝑡𝑒𝑠
𝟑 (𝟖−𝟑)!𝟑! 𝟓!𝟑!
✓ If one engineer is selected at random from the company, find the probability that,
A) he has only a bachelor’s degree.
B) he has a master’s degree, given that he is over 40.
11/10/2023
C) he is under 30, given thatStatistics
he has only a bachelor’s degree.
for Economics_AAU 59
END !
11/10/2023 Statistics for Economics_AAU 60
Special Probability Distributions & Densities
Department of Economics
1
Some Special Discrete Probability Distributions
There are many special discrete probability distributions that have different applications and
𝒙 𝟏 𝒙
𝒙 𝟏 𝒙
Alternatively, let , then
Bernoulli Trial is an experiment only with two possible outcomes; and
(Boy or Girl, dead or alive, adopt or not adopt, fail or pass etc.).
The Bernoulli distribution arises when the following three conditions are satisfied.
Each trial result in just two outcomes.
The probability of success, is the same for each trial.
The trials are independent.
Example 1: Consider tossing of a coin and let X = an event of observing head.
Outcomes X P( 𝒊)
Tail 0
Head 1
From the outcomes, we can calculate,
𝟏 𝟏 𝟏
. And,
𝟎 𝟏 𝟎
𝟏𝒙 𝟏 𝒙
𝟏 𝟏 𝟓
Solution: Here and 𝒊 𝟔
𝟔 𝟔 𝟔
𝟏
and,
𝟔
𝟏 𝟓 𝟓
V
𝟔 𝟔 𝟑𝟔
ii. Compute the first four moment about the mean for this Bernoulli Distribution.
Questions
I. How many possible outcomes can there be for a Bernoulli trial ?
II. In Bernoulli trial, if the probability of success is , what will be the
probability of failure?
III. In a toss of a single fair die, compute the mean and variance of the distribution if an
even number occurs.
The Binomial Distribution
𝒏 𝒙 𝒏 𝒙
𝒊
𝒙
𝒏
where: 𝒙
denotes a combination; is the number of trials; is the random variable
defined as the number of successes; and is the probability of a success on each trial.
A Binomial Experiment is any experiment that can be considered as a sequence of
n trials in which,
i. The number of trials 'n' is finite and determined before the experiment
begins,
ii. Each trial results in two mutually disjoint outcomes, termed as success
and failure,
iii. The random variable is the number of successes in a fixed number of trials.
iv. The result of any trial is independent of the results of all other trials,
v. The probability of success and failure does not change from trial to trial.
Illustration: Lets consider a toss of a coin two times.
Possible outcomes
Now lets consider getting one head as the success. Count the number of successes in each
possible outcome. X = number of heads, what is the probability that ; ; and
?
X= No. of heads P( 𝒊
0
1
2
ii)
𝟑 𝟑 𝟑
𝟑 𝟏
Thus, .
𝟖 𝟖
𝟐
• Exercise 1: A and B play a game in which A’s chance of wining is .
𝟑
• In a series of 8 games what is the probability that A will win at least 6 games?
Mean and Variance: Let X be a binomially distributed random variable with parameter p and
n, denoted by .
𝒏 𝒏 𝒏 𝒙 𝒏 𝒙
Then, 𝒙 𝟎 𝒊 𝒊 𝒙 𝟎 𝒊 𝒙
𝒏
𝒙 𝒏 𝒙
𝒊
𝒙 𝟎
𝟐 𝟐
Example 1: If the probability of defective bolt is 0.1. Find:
(a) the mean and standard deviation for the distribution of defective bolts in a total of
500 bolts,
(b) the distribution of the random variable, and
(c) The first four moments about the mean?
Solution:
a) ,
Then,
Therefore,
𝟓𝟎𝟎 𝒙 𝟓𝟎𝟎 𝒙
𝒙
𝒊
The Hyper Geometric Distribution
For the binomial distribution to be applied, the probability of a success must stay the same
For example, the probability of guessing the correct answer to a true/false question is 0.50.
Most sampling, however, is done without replacement. Thus, if the population is small, the
probability of a success will change for each observation. When the population is finite and the
sampling is done without replacement, so that the events are stochastically dependent,
2) If the size of the sample n is more than 5% of the size of the population N.
𝑺 𝑵 𝑺
𝒙 𝒏 𝒙
𝑵
𝒏 , for x = 0,1,..n
Where: N is the size of the population; S is the number of successes in the
population; x is the number of successes in the sample: It may be 0, 1, 2, 3, . . . ; and
n is the size of the sample or the number of trials.
In summary, a hypergeometric probability distribution has these characteristics:
An outcome on each trial of an experiment is classified into one of two mutually
exclusive categories: a success or a failure.
The random variable is the number of successes in a fixed number of trials.
The trials are not independent.
𝒏
We assume that we sample from a finite population without replacement and
𝑵
> 0.05. So, the probability of a success changes for each trial.
Example: From a basket of 10 items, containing 3 defectives a sample of 4 items is drawn at
random, without replacement.
Let the random variable X denote the number of defective items in the sample space. Find,
a) Probability of obtaining exactly 2, defective items,
b) Obtain
Solution:
Given ; ;
𝑺 𝑵 𝑺 𝟑 𝟏𝟎 𝟑 𝟑 𝟕
𝒙 𝒏 𝒙 𝟐 𝟒 𝟐 𝟐 𝟐 𝟑
𝑵 𝟏𝟎 𝟏𝟎 𝟏𝟎
𝒏 𝟒 𝟒
𝟑 𝟕 𝟑 𝟕
𝟎 𝟒 𝟏 𝟑
𝟏𝟎 𝟏𝟎
𝟒 𝟒
Example 2: The college of Business and Economics have 50 academic staff. 40 of
them are in the department of Economics. The dean office wants to form a
committee of 5 persons for overseeing the operation of the Staff Lounge. What is the
probability that 4 of the selected committees will be from the department of
Economics?
Solution:
Given ; ;
Then, = = =
𝒏 𝒏
𝒊 𝒊
𝒙 𝟎 𝒙 𝟎
𝑺
Then, if we denote
𝑵
𝟐 𝟐
𝒏𝑺
Knowing that and upon longways of simplification of the expressions are;
𝑵
𝟐
The Poisson Distribution
The Poisson probability distribution describes the number of times some event occurs during
a specified interval. Examples of an interval may be time, distance, area, or volume.
The distribution is based on two assumptions:
1) The probability is proportional to the length of the interval
2) The intervals are independent
To put it another way, the longer the interval, the larger the probability, and the number
of occurrences in one interval does not affect the other intervals. This distribution is a
limiting form of the binomial distribution when the probability of a success is very small
and n is large.
It is often referred to as the “law of improbable events,” meaning that the probability, , of a
particular event’s happening is quite small.
A random variable X is defined to have Poisson distribution if the discrete density function of
µ𝒙 𝒆 µ
is given by: , where: μ (mu) = p is the mean number of occurrences (successes)
𝒙!
If the number of lost bags per flight follows a Poisson distribution with , we can
µ𝒙 𝒆 µ
compute the various probability by the formula:
𝒙!
In others words, 97.05 percent of the flights will have no lost baggage.
The probability of exactly one lost bag:
𝟏 𝟎.𝟎𝟑
Thus, we would expect to find exactly one lost bag on 2.9 percent of the flights.
Questions
1) In a Poisson distribution .
a) What is the probability that ?
b) What is the probability that ?
2) In a Poisson distribution .
a) What is the probability that ?
b) What is the probability that ?
c) What is the probability that ?
Mean and Variance: The mean of the Poisson Distribution is given by:
𝒏 𝒏 µ𝒙 𝒆 µ µ 𝒏 µ𝒙
𝒙 𝟎 𝒊 𝒊 𝒙 𝟎 𝒊 𝒙! 𝒙 𝟎 𝒊 𝒙!
𝟐 𝟐
𝟐
Given µ and finding after longways of simplification we obtain,
µ . Therefore,
Example 1: The probability that a check cashed by a bank will bounce is 0.0003, and 10, 000
checks are cashed, then what will be the mean number of bad checks its variance?
Solution:
The variance will also be the same as the mean i.e
Uniform Density (UD) & The Normal Distribution
Uniform Density is the simplest distribution for a continuous random variable.
𝟏
𝒃 𝒂 Where the parameters a and b satisfy
Properties of Uniform Distribution
1. A uniformly distributed random variable has a constant pdf over the interval of
definition.
3) A uniformly distributed random variable represents the continuous example to equally likely
outcomes.
That is for any sub interval where , is the same for all
interval having the same length. That is,
𝒅 𝒅 𝒄
𝒄 𝒃 𝒂
𝟐 𝟐 (𝒃 𝒂)𝟐
𝟏𝟐
Example 1: A point is chosen at random on the line segment . What is the probability that
𝟑
the chosen point lies between 1 and ?.
𝟐
/
Normal Probability Distribution
Most numerical values of a random variable are spread around the center, and
greater the distance a numerical value has from the center, the fewer numerical
values have that specific value.
A frequency distribution of values of random variable observed in nature which
follows this pattern is approximately bell shaped.
A special case of distribution of measurements is called a normal curve (or
distribution).
If a population of numerical values follows a normal curve and X is the randomly
selected numerical value from the population, then X is said to be normal random
variable, which has a normal probability distribution.
The random variable X is said to have a normal (or Gaussian) distribution with parameter
(mean) and (standard deviation) if its pdf is given by:
𝟏 𝒙 µ 𝟐
𝟏 ( 𝟐 )
𝟐 𝝈 for
𝝈 𝟐𝝅
4) Since the normal curve is symmetric, the mean, median, and mode for the normal
distribution are equal because the highest value of the probability density function occurs
when value of a random variable,
5) The two tails of the normal curve extend to infinity in both directions and theoretically never
touch the horizontal axis.
6) The mean of the normal distribution may be negative, zero, or positive.
7) The mean µ determines the central location of the normal distribution, while standard
deviation σ determines its spread.
The larger the value of the standard deviation σ, the wider and flatter is the normal curve,
thus showing more variability in the data.
Thus standard deviation σ determines the range of values that any random variable is
likely to assume.
8) The area under the normal curve represents probabilities for the normal random variable, and
therefore, the total area under the curve for the normal probability distribution is 1.
Figure a: Characteristics of a Normal Distribution
Standard Normal Probability Distribution
There is not just one normal probability distribution, but rather a “family” of them. In
application, it is necessary that a normal random variable X is standardized by
expressing its value as the number of σ it lies to the left or right of its µ.
The standardized normal random variable, z (also called , or
normal variate) is defined as:
OR equivalently, .
A measures the number of standard deviations that a value of the random
variable X fall from the mean.
is normal probability distribution with mean zero and standard deviation one.
From the formula, we may conclude that
i) If x is less than the mean (µ), the value of z is negative.
ii) If x is more than the mean (µ), the value of z is positive.
iii) When , the value of
Figure b: Standard Normal Distribution
Transforming measurements to standard normal deviates changes the scale. The conversions
are also shown in the previous graph. For example, is converted to a z value of 1.00.
Likewise, is transformed to a z value of 2.00.
Note that the center of the z distribution is zero, indicating no deviation from the mean, .
The Empirical Rule
It states that if a random variable X is normally distributed, then:
Approximately 68% of the observations will lie within plus and minus 1 standard
deviation of the mean .
About 95% of the observations will lie within plus and minus 2 standard deviations of
the mean .
Practically all, or 99.7% of the observations, will lie within plus and minus 3 standard
deviations of the mean .
Area Under the Normal Curve
The area under the standard normal distribution between the mean and a
specified positive value of z, say 𝟎 is the probability 0≤z≤ 𝟎) and can be read off
directly from STANDARD NORMAL (Z) TABLES.
For example, area between is the proportion of the area under the curve
which lies between the vertical lines erected at two points along the x–axis.
The standard normal distribution is very useful for determining probabilities for
any normally distributed random variable.
Basic Procedures in Finding Area
1) Find the z value for a particular value of the random variable based on the mean
and standard deviation of its distribution.
2) Use the z table value to find various probabilities(area under the curve).
Table a: Z-Score of a standard normal probability distribution, 0≤z≤ )
NOTE: The table above is the area to the one side of the center (to the Right Side)
We can also have cumulative probability.
Example 1: Suppose the weekly income of taxi drivers follows the normal
probability distribution with a mean of $1,000 and a standard deviation of $100. What is
the z value of income for
1) a driver who earns $1,100 per week?
2) a driver who earns $900 per week?
Solution: Given, ; 𝟏 ; and 𝟐
𝒙 µ 𝟏,𝟏𝟎𝟎 𝟏𝟎𝟎𝟎
𝟏 𝝈 𝟏𝟎𝟎
𝒙 µ 𝟗𝟎𝟎 𝟏𝟎𝟎𝟎
𝟐 𝝈 𝟏𝟎𝟎
Implication:
The z of 1.00 indicates that a weekly income of $1,100 is 1 standard deviation above the
mean, and a z of −1.00 shows that a $900 income is 1 standard deviation below the mean.
Example 2: The lifetimes of certain kinds of electronic devices have a mean of 300 hours and
standard deviation of 25 hours. Assuming that the distribution of these lifetimes, which are
measured to the nearest hour, can be approximated closely with a normal curve,
Find the probability that any one of these electronic devices will have a lifetime of
more than 350 hours.
What percentage will have lifetimes of 300 hours or less?
What percentage will have lifetimes from 220 or 260 hours?
Solution: Given, ; Then,
Thus,
Therefore, the required percentage is, 0.5000 × 100 = 50%.
c) Given ; ; 𝟏 ; and 𝟐 .
Then,
𝒙 µ 𝟐𝟐𝟎 𝟑𝟎𝟎
𝟏 𝝈 𝟐𝟓
𝒙 µ 𝟐𝟔𝟎 𝟑𝟎𝟎
𝟐 𝝈 𝟐𝟓
From the normal table, we have
(Area
and ) and
As 8 percent of the items are above 64, therefore area to the right of the ordinate at 64 is
0.08. Area to the left of the ordinate at up to mean ordinate is
and the value of z corresponding to this area is 1.4 (obtained from the table).
𝒙 µ 𝟔𝟒 µ
Hence,
𝝈 𝝈
From these two equations, we get and . Thus, mean of the distribution
is 50 and standard deviation 10.
a) What percentage of people taking the test score are between 400 and 500?
b) Suppose someone received a score of 630. What percentage of the people taking the test
c) If the particular college will not admit any one scoring below 480, what percentage of the
a) The area under the normal curve between and z = 0.5 is 0.1915. The probability that
the score falls between 400 and 500 is =
So the percentage of the people taking the test score
between 400 and 500 is 38.30 percent.
𝒙 µ 𝟔𝟑𝟎 𝟒𝟓𝟎
b) Given ,
𝝈 𝟏𝟎𝟎
Thus, the percentage of people who score more than 480 and are acceptable to the
university is 38.21 percent.
Gamma and Beta Distributions
Gamma Distribution: The PDF for the gamma (or Erlang) distribution is:
𝝁(𝝁𝒙)𝒏 𝟏 𝒆 𝝁𝒙
, ,
𝒏 𝟏 !
𝒎 𝟏 𝒏 𝟏
, ;
Applications:
Beta distribution is commonly used to describe the random variable whose possible
values lie in a restricted interval of numbers.
A typical use of this distribution is found in the use of program Evaluation and
Review Technique (PERT) and Critical Path Method(CPM) where activity times are
estimated within a specific range.
Highly applicable in Project Management. For instance, the beta distribution
technique in Project management is used to identify uncertainty in the estimated
project time.
The Normal Approximation of Binomial Distribution
In the first section of this chapter we discussed the binomial probability distribution,
We can use the normal distribution (a continuous distribution) as a substitute for a binomial
The End.
Thanks
Joint and Conditional Probability Distribution
Department of Economics
1
Introduction: Joint and Conditional Probability Distribution
dimensional case).
random variables. That is., Education and Earnings; Height and Longevity;
Attendance and Learning Outcomes; Sex-ratios and Areas Under Rice Cultivation;
income and expenditure; rainfall(R) and average temperature (T) Genetic Make-Up
(3) a silver full-size sedan; the probability distribution for purchasing would
not be the same for women in their 20s, 30s, and 50s.
Thus, it is important that probability models reflect the joint effect of variables on
probabilities.
Joint Probability: is probability of two or more independent events occurring
together or in succession.
The joint probability of two or more independent events is equal to the
product of their marginal probabilities.
In particular, if A and B are independent events, the probability that both A
and B will occur is given by:
Example 1: Suppose we toss a coin twice. The probability that in both cases the
coin will turn up head is given by:
In general, if X and Y are two random variables, the probability distribution that
It is denoted as:
𝟏 𝟐 and 𝟏 𝟐 respectively.
Then, the Joint Probability Mass Function of is: 𝒊 𝒊
𝒊 𝒊 .
The values give the probability/likelihood that outcomes X and Y occur at the same
time.
Example 1: CDs are covered with a rectangular plastic. Measurements for the
length and width are rounded to the nearest mm (so they are discrete). Let X
denote the length and Y denote the width. The possible values of X are 129, 130,
and 131 mm while the possible values of Y are 15 and 16 mm (Thus, both X and Y
are discrete). Thus, there are six possible areas (pairs of X and Y) of the plastic
cover. The probability for each pair is shown in the following table:
X: length
129 130 131
15 0.12 0.42 0.06
Y: Width
16 0.08 0.28 0.04
From the table, the sum of all the probabilities is 1.0, the combination with the highest
probability is (130, 15), the combination with the lowest probability is (131, 16).
The joint probability mass function is the function 𝑿𝒀 𝑿𝒀 . For
example, we have 𝑿𝒀 .
Recall: A marginal or unconditional probability is the simple probability of the occurrence
of an event.
Which is the probability of the occurrence of a single event i.e.,
.
𝟏
Examples: The probability a card drawn is red is and the probability of a
𝟐
𝟏𝟑 𝟏
card drawn is a heart is .
𝟓𝟐 𝟒
Thus, if we are given a joint PMF for X and Y, we can obtain the individual
probability distribution for X or for Y (and these are called the Marginal
Probability Distributions).
Given the previous example and find the probability that a CD cover has length of 129 mm
(i.e. ). Then, what is the probability distribution of X?
Solution:
The same way, the probability distribution for Y appears in the raw totals, given
as: Y 15 16
P( 0.60 0.40
Because the PMF for X and Y appear in the margins of the table (i.e. column and
row totals), they are often stated to as the Marginal Distributions for X and Y.
We can also compute the Expected Values (i.e. and ) and variance
of a single random variable from the joint probability distribution function.
Example 2: A fair coin is tossed three times independently: let X denote the number of
heads on the first toss and Y denote the total number of heads. Then, find the joint
probability mass function of X and Y .
Solution: the joint distribution of can be summarized in the following table:
X|Y 0 1 2 3
0 1/8 2/8 1/8 0
1 0 1/8 2/8 1/8
Suppose that we wish to find the PMF of Y from the joint PMF of X and Y in the previous
example:
𝒚 )
𝒚 )
In general, to find the frequency function of Y, we simply sum down the
ii. the sum of the joint probabilities over all possible pairs of values must be 1,
i.e. 𝑿 𝒀
𝑿 𝒙 𝒙
𝟐 𝟐
𝒙 𝒙 𝑿 𝒙
𝒚 𝒚 𝒚
𝟐 𝟐
𝒚 𝒚 𝒚 𝒚
Exercise 1: A basket contains two Red, three Black, and four White balls. Let X be the number of chosen
black balls and Y be number of chosen red balls. If two balls are randomly selected from the basket(without
replacement),
A) Find the joint probability distribution function
B) The variance of X and Y.
Exercise 2: Given the joint PMF of two random variables as follow,
𝑷(𝒙, 𝒚) 0 10 20
10 1/5 1/10 1/5
25 5/100 3/20 3/10
If X and Y are continuous random variables and is the value of their joint
probability density at , the function given by:
𝒇(𝒙,𝒚)
distribution/density of Y given is defined as: if
𝒇(𝒙)
This definition holds for both discrete and continuous random variables. You can also notice
the similarity between the conditional probability and conditional probability distribution.
Example:
Age Group (X)
Purchase Decision 1 2 3
(Y) (16 to 20) (26 to 46) (46 to 65)
1(buy) 0.10 0.20 0.10 0.40
2(not buy) 0.25 0.25 0.10 0.60
0.35 0.45 0.20 1.00
Statistical Independence: The condition of statistical independence arises when the
occurrence of one event has no effect upon the probability of occurrence of any other event.
For example, in a fair coin toss the outcome of each toss is an event that is statistically
independent of the outcomes of every other toss of the coin.
Note: The probability computations under statistical dependent and independent events
is very different.
Under Statistical dependence
and
OR
𝑷(𝑨∩𝑩) 𝑷(𝑨∩𝑩)
and
𝑷(𝑩) 𝑷(𝑨)
When the information that the random variable X takes a particular value x is irrelevant to the
determination of the probability that another random variable Y takes, we say that Y is
independent of X.
Formally, two random variables, X and Y, are said to be independent if and only if any one of
the following three conditions holds:
for
, for
, for
Exercise 1: Given the joint PDFs in the previous exercises, compute the conditional
probability density functions for X given Y and for Y given X.
Expectations: Mathematical Expectation
If X and Y are two discrete random variables with joint PDF, then:
, 𝒙 𝒚
Given 𝒚
Example: Given the PMF in the form of table as below, find, , and
X|Y 0 1 2 3
0 1/8 2/6 1/8 0 4/8
1 0 2/8 1/8 1/8 4/8
1/8 3/8 3/8 1/8 1
Solution:
𝟒 𝟒
𝒙 𝟖 𝟖
𝒚
𝟏 𝟑 𝟑 𝟏 𝟏𝟐
𝟖 𝟖 𝟖 𝟖 𝟖
𝟏 𝟐
𝒙 𝒚 𝟖 𝟖
𝟐
The Variance,
𝟐 𝟐 𝟐 𝟐
𝟐 𝟐 𝟐 𝟐
𝟐 𝟐
𝒙 𝒙 , for continuous random variable.
𝟐
Note that the variance for Y, 𝒚 is also expressed in the same way, replacing Y in place of X
in the expressions above.
Exercise: Compute the variance for X and Y for the previous example. Answer:
𝟐 𝟐
𝒙 𝒚
Covariance and Correlation
An important measures when we have two random variables is that of covariation and
correlation of X and Y. The relationship between two variables can be represented by a
covariance and a correlation, which then suggests that there may be a relationship that differs
from 0.
Covariance: measures the amount of linear dependence between two random variables.
Is a measure of the joint variability of two variables.
A positive covariance indicates that two random variables move in the same direction,
while a negative covariance indicates they move in opposite directions.
Interpreting the magnitude of a covariance can be a little tricky.
It is defined as follows:
𝑿 𝒀 𝑿 𝒀
𝒚 𝒙 𝑿 𝒀 , in the case of discrete.
𝒚 𝒙 𝑿 𝒀
Correlation: the degree of association between a dependent (response) variable Y and two or
more independent variables.
is an alternative measure of dependence between X and Y that solves the “units” problem
of the covariance.
Specifically, the correlation between X and Y is the covariance between X and Y divided
by their standard deviations, denoted by or 𝑿,𝒚 is:
𝑪𝑶𝑽 𝑿,𝒀 𝑪𝑶𝑽(𝑿,𝒀)
or 𝑿,𝒚 𝑽𝒂𝒓 𝒙 ∗𝑽𝒂𝒓(𝒀) 𝝈𝑿 𝝈𝒀
This is called the coefficient of correlation, which measures linear relationship between X
and Y. Thus, Covariance measures the joint variability of two random variables while
Correlation measures the strength/the degree to which variables are moving together.
Properties of the coefficient of correlation
𝑿,𝒚
and
𝒙,𝒚 𝒚,𝒙
Exercise 1: Find the covariance and coefficient of correlation of our example under the mean
for the discrete.
𝟏 𝟏
Answer: 𝒙,𝒚 𝒚,𝒙
𝟒 𝟑
𝟑 𝟐 𝟒
𝒙 𝒚 𝟐 𝒙 𝒚
Exercise 2: Given 𝑵 with ; , and ; and
𝒏
𝟐 ; 𝟐 ; ;
𝒙 𝒚
𝟕 𝟕
𝒙,𝒚 𝒚,𝒙
𝟓𝟒 𝟓𝟒
𝟕 𝟒𝟗 𝟑𝟒𝟑
∗
𝟏𝟖 𝟏𝟔𝟐 𝟐𝟗𝟏𝟔
The covariance ends up negative because of the restriction given by . So that when X
increases, Y must go down and thus, the negative relationship between the two random variables.
Conditional Expectation
Conditional expectation of X given is defined as:
𝒙 , if discrete.
, if continuous.
Similarly, the conditional expectation of Y given is defined as:
𝒚 , if discrete.
, if continuous.
Note that: The conditional expectation of Y given turns out to be the function of X. That
is, or is called the regression function of Y on X.
It tells us how the mean of Y varies with changes in X.
Example: In our earlier example given as follows
X|Y 0 1 2 3
0 1/8 2/6 1/8 0 4/8
1 0 2/8 1/8 1/8 4/8
1/8 3/8 3/8 1/8 1
Find,
Solution: a) Conditional probability distribution of X
Y
0
1
2
3
a) ( | = ), Conditional Probability distribution of Y
X ( | = ) ( | = ) ( | = ) ( | = )
0
𝟏 𝟏 𝟏
𝒚 𝟒 𝟐 𝟒
X
0 1
1 2
𝒙 𝒚
Proof:
Note: If two random variables are independent, then they have zero correlation. However, the
converse is not necessarily true. That is, two random variables may have zero correlation; it
does not necessarily mean that they are independent.
Example: Let the joint PMF of two random variables X and Y be given as follows:
X|Y -1 0 1
-1 0 1/4 0 1/4
0 1/4 0 1/4 1/2
1 0 1/4 0 1/4
1/4 1/2 1/4 1
; and , therefore
However, is not equal to
.
Note also that is not equal to ; but X and Y are dependent.
Y Y
-1 0.5 -1 1/4
0 0 0 1/2
1 0.5 1 1/4
1 1
The sum of the conditional distribution should add up to unity. Similarly, the sum
of the marginal distribution also adds up to one. If X and Y are independent
for all X, but in the above example this does not hold, therefore,
the two variables are not independent.
1
5-2
Sampling Distributions
• The sampling distribution of a statistic is the probability
distribution of all possible values the statistic may assume,
when computed from random samples of the same size,
drawn from a specified population.
• The sampling distribution of X is the probability
distribution of all possible values the random variable X may
assume when a sample of size n is taken from a specified
population.
5-5
Sampling Distributions
• The concept of sampling distribution can be related to the
various probability distributions.
• Probability distributions are the theoretical distributions of
random variables that are used to describe characteristics of
populations or processes under certain specified conditions.
• That is, probability distributions are helpful in determining the
probabilities of outcomes of random variables when
populations or processes that generate these outcomes satisfy
certain conditions.
5-6
Sampling Distributions
• When the population has a normal distribution, then the
phenomenon that describes the normal probability distribution
provides a useful description of the distribution of population
values.
• When the mean values obtained from samples are distributed
normally, it implies that this distribution is useful for describing the
characteristics (properties) of sampling distribution.
• Consequently, these properties, which are also the properties of
sampling distribution, help to frame rules for making statistical
inferences about a population on the basis of a single sample drawn
from it, that is, without even repeating the sampling process.
5-7
Sampling Distributions
• The sampling distribution of a sample statistic also helps in
describing the extent of error associated with an estimate of the
value of population parameters.
5-8
Cont’d
• In such distributions the arithmetic mean represents the
average of all possible sample means or the ‘mean of means’
denoted by X ; the standard deviation which measures the
variability among all possible values of the sample values, is
considered as a good approximation of the population’s standard
deviations σ.
• To estimate σ of the population to greater accuracy the formula
5-10
Cont’d
• where n is the size of sample. The new value n – 1 in the
denominator results into higher value of s than the observed value
s of the sample. Here n – 1 is also known as degree of freedom.
• The number of degrees of freedom, df = n – 1 indicate the number
of values that are free to vary in a random sample.
• Sampling distribution is the distribution of all possible values
of a statistic from all the distinct possible samples of equal size
drawn from a population or a process.
5-11
P(X)
0.1
variance.
0.10
P(X)
0.05
0.00
1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 6.0 6.5 7.0 7.5 8.0
X
5-15
Relationships between Population Parameters and the Sampling
Distribution of the Sample Mean
The expected value of the sample mean is equal to the population mean:
E( X )
X X
The variance of the sample mean is equal to the population variance divided by
the sample size:
2
V (X ) 2
X
X
n
The standard deviation of the sample mean, known as the standard error of
the mean, is equal to the population standard deviation divided by the square
root of the sample size:
SD ( X ) X
X
n
5-16
Sampling from a Normal Population
When sampling from a normal population with mean and standard
deviation , the sample mean, X, has a normal sampling
distribution:
2
X ~ N (, )
n
This means that, as the sample Sampling Distribution of the Sample Mean
Sampling Distribution: n = 2
population mean, but becomes 0.1
Normal population
Normal population
more compactly distributed 0.0
around that population mean
5-17
The Central Limit Theorem
• If population is not normally distributed, then we make use of the
central limit theorem to describe the random nature of the
sample mean for large samples without knowledge of the
population distribution.
• The Central Limit Theorem states that when the random
samples of observations are drawn from a non-normal population
with finite mean µ and standard deviation σ, and as the sample
size n is increased, the sampling distribution of sample mean x
is approximately normally distributed, with mean and standard
deviation as:
5-18
Cont’d
• Regardless of its shape, the sampling distribution of sample
mean x always has a mean identical to the sampled
population, i.e. µx = µ and standard deviation σx = σ / x . This
implies that the spread of the distribution of sample means is
considerably less than the spread of the sampled population.
• The central limit theorem is useful in statistical inference.
• When the sample size is sufficiently large, estimations such as
‘average’ or ‘proportion’ that are used to make inferences about
population parameters are expected to have sampling distribution
that is approximately normal.
5-19
Cont’d
• The behavior of these estimations can be described in repeated sampling and
are used to evaluate the probability of observing certain sample results using
the normal distribution as follows:
n=5
When sampling from a population 0.25
P(X)
0.10
P(X)
0.1
(n >30). Large n
0.4
0.3
f(X)
For “large enough” n: X ~ N(, / n)
0.2
2
0.1
0.0
-
X
5-21
The Central Limit Theorem Applies to Sampling Distributions from Any Population
Population
n=2
n = 30
X X X X
5-22
The Central Limit Theorem
Example: Mercury makes a 2.4 liter V-6 engine, the Laser XRi, used in speedboats.
The company’s engineers believe the engine delivers an average power of 220
horsepower and that the standard deviation of power delivered is 15 HP. A
potential buyer intends to sample 100 engines (each engine is to be run a single
time). What is the probability that the sample mean will be less than 217 HP?
X 217
P ( X 217) P
n n
217 220 217 220
P Z P Z
15 15
100 10
P ( Z 2) 0.0228
5-23
For this purpose, the value of sample mean x is first converted into a value z on
the standard normal distribution to know how any single mean value deviates
from the mean x of sample mean values, by using the formula
5-25
5-26
5-27
Student’s t Distribution
Degrees of Freedom
Degrees of freedom: The number of unrestricted chances for
variation in the measurement being made.
The devisor (n–1) in the formula for the sample variance s2 is called
number of degrees of freedom (df) associated with s2.
The number of degrees of freedom refers to the number of unrestricted
chances for variation in the measurement being made, i.e. number of
independent squared deviations in s2 that are available for estimating σ2.
In other words, it refers to the number of values that are free to vary in a
random sample.
The shape of t-distribution varies with degrees of freedom. Obviously
more is the sample size n, higher is the degrees of freedom.
5-29
Degrees of Freedom
Example: The mean length of life of a certain cutting tool is 41.5
hours with a standard deviation of 2.5 hours. What is the
probability that a simple random sample of size 50 drawn from
this population will have a mean between 40.5 hours and 42
hours?
Solution: We are given the following information µ = 41.5 hours,
σ = 2.5 hours, and n = 50
It is required to find the probability that the mean length of life, x
, of the cutting tool lies between 40.5 hours and 42 hours, that is,
P(40.5 ≤ x ≤ 42).
5-30
Cont’d
5-31
5-32
Degrees of Freedom
If only two data points and the sample mean are known:
x 12 14 x x 4
3
x= 14
n 4
12 14 x x 4 56
3
5-34
Degrees of Freedom
The number of degrees of freedom is equal to the total number of
measurements (these are not always raw data points), less the total
number of restrictions on the measurements. A restriction is a
quantity computed from the measurements.
The sample mean is a restriction on the sample measurements, so after
calculating the sample mean there are only (n-1) degrees of freedom
remaining with which to calculate the sample variance. The sample
variance is based on only (n-1) free data points:
s 2
( x x ) 2
(n 1)
Cont’d
5-35
Sample 93 97 60 72 96 83 59 66 88 53
Point
Solution: Choose the means of the corresponding sample points. These are: 83.6,
69.33, and 70.5.
Solution: df = 10 – 3 = 7.
n=2, p = 0.3
0 .4
P(X)
0 .2
number of trials, n.
0 .0
0 1 2
n=10,p=0.3
0.3
X
Sample proportion: pˆ 0.2
n
P(X)
0.1
0.0
0 1 2 3 4 5 6 7 8 9 10
X
P(X)
deviation p (1 p )
0.1
n 0.0
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 X
0 1 2 3 4 5 6 7 8 9 10 11 12 13 1415
15 1515 15 15 15 15 15 15 1515 15 15 15 1515 ^p
Sample Proportion
5-39
• Example: In recent years, convertible sports coupes have become very popular in Japan.
Toyota is currently shipping Celicas to Los Angeles, where a customizer does a roof lift and
ships them back to Japan. Suppose that 25% of all Japanese in a given income and lifestyle
category are interested in buying Celica convertibles. A random sample of 100 Japanese
consumers in the category of interest is to be selected. What is the probability that at least
20% of those in the sample will express an interest in a Celica convertible?
n 100
p 0.25
p p .20 p
P ( p > 0.20 ) P
p (1 p )
>
p (1 p )
np (100 )( 0.25) 25 E ( p )
n n
p (1 p )
(.25)(.75)
0.001875 V ( p ) .20 .25 > .05
P z > P z
n 100
(.25)(.75)
.0433
p (1 p ) 100
0.001875 0.04330127 SD ( p ) P ( z > 1.15) 0.8749
n
5-40
40
•
ESTIMATION
: the sample
variance is unbiased & consistent estimator of the population variance.
It is relatively efficient as compared to other estimators.
Its equal root, the sample standard deviation, is generally used as an
estimator of the population standard deviation is also relatively efficient.
𝟐
𝒊
It can be computed as:
: is an unbiased, consistent, and relatively
efficient estimator of the population proportion.
Due to this fact, it is generally the preferred estimator of the
population proportion. It is calculated by taking elements in the
sample that have the same characteristics.
Example: Price- earnings ratios for a random sample of ten shares
traded in Addis Ababa stock Exchange on December 27, 2011 were:
10; 16; 5; 10; 12; 8; 4; 6; 5; 4
Find point estimates of the population mean, variance and standard
deviation and Proportion of shares in the population for which the
price earnings ratio exceeded 8.5
Solution
To find the first three of these sample quantities, we show the
calculations in tabular form as:
i 1 2 3 4 5 6 7 8 9 10 Total
Xi 10 16 5 10 12 8 4 6 5 4 80
Where 𝜶
𝟐
is the value of standard normal variable that is exceeded
with a probability of or 𝜶
𝟐
is Z value providing an area of in
the upper tail of the standard normal probability distribution.
Solution: given
Required: 𝜶
𝟐
𝜶
𝟐
, 1st find the standard error
𝑿
of the mean i.e.,
Then we have that = level of significance = 0.05 since = 0.95, so
Therefore, the confidence interval estimation for the population mean ( ) is:
In this case we may say that “we are 95% confident that the
population mean lies within 46.08 and 53.92.”
This statement does not mean that the chance is 0.95 that the
population mean of all the random variables falls within the interval
established from this one sample.
Instead, it means that if we select many random samples of the same
size and if we compute a confidence interval for each of these
samples, then in about 95 percent of these cases, the population mean
will lie within that interval.
Required:
Note that:
A narrower confidence interval is more precise
Larger samples give more precise estimates
Small variance leads to more precise estimates
Lower confidence coefficients allow us to construct more precise
estimate
Calculating Confidence Interval Estimate of a Population Mean:
Normal Population with Unknown
When the population standard deviation is unknown, we use the
sample standard deviation, as an estimate of . The sample
standard deviation is given by:
Case 2: when the sample size is small and unknown (A sample size
is small when n < =30)
Confidence interval estimate for population mean ( ) is given by:
𝑿
𝜶 𝜶
𝟐 𝟐
Required:
Example 2: In the testing of a new production method, 18 employees were
selected randomly and asked to try the new method. The sample mean
production rate for the 18 employees was 80 parts per hour and the sample
standard deviation was 10 parts per hour. Provide 90% and 95% confidence
intervals for the population mean production rate for the new method,
assuming the population has a normal probability distribution.
Solution: given
Required:
Determining the Sample Size in Estimating Population Mean
Some sampling error will arise because we have not studied the whole
population.
We need some criteria for judging the goodness of the estimate. The
criteria or properties for a good estimator may be different for small
and large samples.
A) Small sample properties
1. Unbiasedness: An estimator is said to be unbiased if the expected
value of the estimator is equal to the true population parameter, i.e. if
.
2. Consistency: An estimator is consistence if it satisfies the following
conditions:
i. The estimator must be asymmetrically unbiased
ii. The variance of the estimator must approach to zero as n approaches
infinity, i.e.,
1. Asymptotic Efficiency: An estimator is asymptotically efficient if:
i. The estimator is consistent
ii. The estimator has a smaller asymptotic variance as compared to any
consistent estimator.
Lecture VII
HYPOTHESIS TESTING
INTRODUCTION
When a sample is drawn from a population, the evidence
obtained can be used to make inferential statements about the
characteristics of the population. As we have seen, one
possibility is to estimate the unknown population parameters
through the calculation of point estimates or confidence
intervals. Alternatively, the sample information can be
employed to assess the validity of some conjecture, or
hypothesis, that an investigator has formed about the
population. A hypothesis is formed about some population,
and conclusions about the merits of this hypothesis are to be
formed on the basis of sample information.
Some terminologies in hypothesis testing
Alternative hypothesis (Ha): the complement of the null
hypothesis or the hypothesis one accepts if he/she rejects the
null hypothesis. The alternative hypothesis can take either of
three forms: greater than, less than or different from.
A hypothesis against which the null hypothesis is tested, and which
will be held to be true if the null is held false.
Simple Hypothesis: A hypothesis that specifies a single value for a
population parameter
Composite Hypothesis: A hypothesis that specifies a range of value
for a population parameter
One Sided Alternative: An alternative hypothesis involving all
possible values of a population parameter on either one side or the
other of (that is, either greater than or less than) the value specified by
a simple null hypothesis.
Two Sided Alternatives: An alternative hypothesis involving all
possible values of a population parameter other than the value
specified by a simple null hypothesis.
Hypothesis Test Decisions: A decision rule is formulated, leading the
investigator to either accept or reject the null hypothesis on the basis
of sample evidence.
Cont’d
One tailed and two tailed test: the test of hypothesis in which
the alternative hypothesis is given by the statement of greater
than or less than is called one tailed test because it is given in
one direction. If the alternative hypothesis is given by the
statement "different from" the test is called a two tailed test.
Step 2: Set the criteria for a decision. To set the criteria for a
decision, we state the level of significance for a test.
Other testing options include the chi-square test and the f-test.
The other possible error, called a Type II error, arises when a false
null hypothesis is accepted.
𝟎
𝜹
𝒏
Z is a standard
random variable.
Example: When a process producing ball is operating correctly, the
weights of the balls have a normal distribution with mean 5 gram and
standard deviation 0.1 gram. An adjustment has been made to the
process, and the plant manager suspects this has raised the mean
weight of ball produced, leaving the standard deviation unchanged. A
random sample of sixteen balls is taken, and their mean weight is
found to be 5.038 gram. Test at significance levels of 0.05 & 0.10
(that is, at 5% and 10% levels) the null hypothesis that the population
mean weight is 5 gram against the alternative that it is bigger.
Solution: given
Since 1.52 does not exceed 1.645, we fail to reject the null hypothesis
at the 5% level of significance; that is, the null hypothesis is accepted
at this significance level.
However, and since 1.52 is bigger than 1.28, the
null hypothesis is rejected at the 10% level of significance. To this
extent, then, there is some evidence in the data to suggest that the true
mean weight exceeds 5 gram.
Obviously the lower the significance level at which a null hypothesis
can be rejected, the greater the doubt cast on its truth.
Rather than testing hypotheses at pre-assigned levels of significance,
investigators often determine the smallest level of significance at
which a null hypothesis can be rejected.
The smallest significance level at which a null hypothesis can be
rejected is called the probability- value, or P-value, of the test.
In the above example we found that Therefore, according
to our decision rule, the null hypothesis is rejected for any significance
level for which is less than 1.52. From the standard normal table
we find that, when is 1.52, is equal to 0.0643= 6.43%. This, then,
is the P– value of the test. The implication is that the null hypothesis
can be rejected at all levels of significance higher than 6.43%.
The appropriate procedure for testing, at a given significance level, the
null hypothesis against the alternative hypothesis
is precisely the same as when the null hypothesis is .
Case II: Tests of the Mean of a Normal Distribution: Population
Variance Known
Suppose we have a random sample of observations from a normal
population with mean and known variance . If the observed
sample mean is , then a test with significance level of of either
null hypothesis
Case III: Tests of the Mean of a Normal Distribution: Population
Variance Known
Suppose we have a random sample of observations from a normal
population with mean and known variance .
If the observed sample mean is , then a test with significance level of
of the null hypothesis
Example: A drill, as part of an assembly line operation, is used to drill
holes in sheet metal. When the drill is functioning properly, the
diameters of these holes have a normal distribution with mean 2
inches and standard deviation 0.06 inch. Periodically, to check that the
drill is functioning properly, the diameters of a random sample of
holes are measured. Assume that the standard deviation does not vary.
A random sample of nine measurements yield mean diameter 1.95
inches. Test the null hypothesis that the population mean is 2 inches
against the alternative that it is not. Use a 5% significance level and
also find the P-value of the test.
Solution: given
Where
Then, since – 2.50 is less than -1.96, the null hypothesis is rejected at
the 5% significance level.
According to the decision rule, the null hypothesis will be rejected for
any significance level for which is bigger than -2.50. From
the standard normal table, we see that when is 2.50, is equal
to 0.006. Hence, = 0.0124. This is the P- value of the test, implying
that the null hypothesis can be rejected against the two-sided
alternative at any level of significance greater than 1.24%. This
certainly casts substantial doubt on the hypothesis that the drill is
functioning correctly.
Tests for the Mean of Large Sample Sizes
Suppose we have a random of n observations from a population with
mean and variance . If the sample size n is large (n>30), the test
procedures developed for the case where the population variance is
known can be employed when it is unknown, replacing by the
observed sample variance . Moreover, these procedures remain
approximately valid even if the population distribution is not normal.
Example: It might be suspected that the firms most likely to attract
take-over bids are those that have been achieving relatively poor
returns. One measure o such performance is through “abnormal
returns,” which average 0 over all firms. A random sample of 88 firms
for which cash tender offers had been made showed abnormal returns
with a mean of -0.0029 and a standard deviation of 0.0169 in the
period from 24 months to 4 months prior to the take-over bids. Test
the null hypothesis that the mean abnormal return for this population
is 0 against the alternative that it is negative.
Solution: given
Example: A real chain knows that, on average, sales in its
stores are 20% higher in December than in November. For the
current year, a random sample of six stores was selected. Their
percentage December sales increases were found to be: 19.2;
18.4; 19.8; 20.2; 20.4 and 19.0. Assuming a normal population
distribution, test the null hypothesis that the true mean
percentage sales increase is 20, against the two-sided
alternative, at the 10% significance level.
Solution: given the data we can find
The decision rule is to reject in favor of if
Where
Thus, since –1.597 lies between - 2.015 and 2.015, the null hypothesis
that the true mean percentage increase is 20 is accepted at the 10%
level. The evidence in the data against this hypothesis is not terribly
strong.
THAKS!!!
END!!!