Chapter IV-2
Chapter IV-2
Distribution
Introduction
So far we have focused on probability distributions for single random
variables(one dimensional case).
However, we are often interested in probability statements concerning
two or more random variables.
Examples:
Education and Earnings; Height and Longevity; Attendance
and Learning Outcomes; Sex-ratios and Areas Under Rice
Cultivation; income and expenditure; rainfall(R) and average
temperature(T) Genetic Make-Up and Disease etc.
In this chapter, we will discuss(both for discrete and continuous) Joint,
Conditional, Marginal Distributions and Independency of random
variables.
4.1 Joint and Marginal Distributions
Business and economic applications of statistics are often concerned about the
relationships between/among variables.
Illustrations:
Products at different quality levels have different prices.
Age groups have different preferences for clothing, for automobiles, and for
music.
The percent returns on two different stocks may tend to be related (positive
relationship; negative relationship).
When we work with probability models for problems involving relationships
between variables, it is important that the effect of these relationships is included
in the Probability Model.
For example, assume that a car dealer is selling the following automobiles: (1) a red
two-door compact, (2) a blue minivan, and (3) a silver full-size sedan; the
probability distribution for purchasing would not be the same for women in their
20s, 30s, and 50s.
Thus, it is important that probability models reflect the joint
effect of variables on probabilities.
Joint Probability: is probability of two or more independent
events occurring together or in succession.
The joint probability of two or more independent events is
equal to the product of their marginal probabilities.
In particular, if A and B are independent events, the
probability that both A and B will occur is given by:
X: length Row
Y: 129 130 131 Sum
Width
15 0.12 0.42 0.06 0.60
16 0.08 0.28 0.04 0.40
ii. Then, the probability distribution for X appears in the column totals, given
as:
X 129 130 131
P( 0.20 0.70 0,10
The same way, the probability distribution for Y appears in the raw
totals, given as:
Y 15 16
P( 0.60 0.40
Because the PMF for X and Y appear in the margins of the table (i.e.
column and row totals), they are often referred to as the Marginal
Distributions for X and Y.
We can also compute the Expected Values (i.e. and ) and variance of a
single random variable from the joint probability distribution
function.
Exercise 1: A basket contains two Red, three Black, and four White
balls.
Let X be the number of chosen black balls and Y be number of chosen
red balls. If two balls are randomly selected from the basket(without
replacement),
A) Find the joint probability distribution function
B) The variance of X and Y.
Exercise 2: Given the joint PMF of two random variables as follow,
0 10 20
10 1/5 1/10 1/5
25 5/100 3/20 3/10
A) Find the probability that .
B) Find the probability that .
C) Find the marginal probability of X, and marginal probability of Y,
D) Compute the expected value and variances of the two random variables.
Exercise 3: Consider two random variable X and Y with joint PMF as given below.
10 15 20
0 1/4 1/6 1/8
1 1/8 1/6 1/6
A) Find the probability that and .
B) Find the probability that .
B) Joint and Marginal Densities: Continuous
A bivariate function can serve as a joint probability density function
of a pair of continuous random variables X and Y if it values, , satisfy the
following two conditions:
i) for and
If X and Y are continuous random variables and is the value of their joint
probability density at , the function given by:
I. and
II. OR
III. and
When the information that the random variable X takes a particular
value x is irrelevant to the determination of the probability that
another random variable Y takes, we say that Y is independent of
X.
Formally, two random variables, X and Y, are said to be
independent if and only if any one of the following three conditions
holds:
i. for
ii. , for
iii. , for
Exercise 1: Given the joint PDFs in the previous exercises,
compute the conditional probability density functions for X given Y
and for Y given X.
4.3 Expectations
Mathematical Expectation
If X and Y are two discrete random variables with joint PDF, then:
,
Given
, for discrete random variable.
The expected value for Y is also given in the same way.
If X and Y are two continuous random variables with joint PDF, then:
Given,
, for continuous random variable.
The expected value for Y is also given in the same way.
If X and Y are two random variables, then,
X|Y 0 1 2 3
0 1/8 2/6 1/8 0 4/8
1 0 2/8 1/8 1/8 4/8
1/8 3/8 3/8 1/8 1
Solution:
The Variance,
Exercise: Compute the variance for X and Y for the previous example.
Answer:
4.3.1 Covariance and Correlation
An important measures when we have two random variables is that
of covariation and correlation of X and Y.
The relationship between two variables can be represented by a
covariance and a correlation, which then suggests that there may
be a relationship that differs from 0.
Covariance: measures the amount of linear dependence between two
random variables.
Is a measure of the joint variability of two variables.
A positive covariance indicates that two random variables move in
the same direction, while a negative covariance indicates they
move in opposite directions.
Interpreting the magnitude of a covariance can be a little tricky.
It is defined as follows:
, in the case of discrete. , in the case of continuous.
Note the following:
1) and
X|Y 0 1 2
0 1/6 2/9 1/36 5/12
1 1/3 1/6 0 1/2
2 1/12 0 0 1/12
Find the correlation
7/12coefficient7/18 1/36Y.
between X and 1
Answer:
;;;
;;;
Note: The covariance ends up negative because of the restriction given by .
So that when X increases, Y must go down and thus, the negative relationship
between the two random variables.
X
0 1
In the1 form of function: 2
4.3.3 Independence and Expectation
Let X and Y be two random variables with probability function , then
1) If X and Y are independent, then
Proof:
Note: If two random variables are independent, then they have zero correlation.
However, the converse is not necessarily true. That is, two random variables may
have zero correlation; it does not necessarily mean that they are independent.
Example: Let the joint PMF of two random variables X and Y be given as follows:
X|Y -1 0 1
-1 0 1/4 0 1/4
0 1/4 0 1/4 1/2
1 0 1/4 0 1/4
1/4 1/2 1/4 1
; and , therefore
However, is not equal to .
Note also that is not equal to ; but X and Y are dependent.
Y Y
-1 0.5 -1 1/4
0 0 0 1/2
1 0.5 1 1/4
1 1
The sum of the conditional distribution should add up to unity.
Similarly, the sum of the marginal distribution also adds up to one.
If X and Y are independent for all X, but in the above example this does
not hold, therefore, the two variables are not independent.
Relationship between Independence and
Uncorrelatedness
Independence: Two random variables are independent when there
is no relationship of any form between the two RVs (X & Y).
Uncorrelatedness: Two random variables are uncorrelated when
there is no relationship of a particular form between the two RVs (X
& Y).
Thus, independent random variables are uncorrelated but
uncorrelated random variables may not be independent.
END!