2 Two Random Variables
2 Two Random Variables
2 Two Random Variables
19
px,y (, ) =
Px,y (, )
px,y ( , )d d
f (, )px,y (, )dd
The discussion of two random variables does involve some new concepts:
reduction to a single variable, conditional probability, and statistical independence. The probability density for a single variable is obtained by integrating
over all possible values of the other variable.
px () =
py () =
px,y (, )d
px,y (, )d
20
Probability
Note that in the expression px (|y), is a variable but y is simply a parameter. px (|y) has all the properties of a probability density function of
a single random variable, . The following picture may be helpful in understanding the connection between the joint probability density px,y (, ) and
the conditional probability density px (|y).
(
(
) =
px,y (, y) d = py ( = y)
c
px ( |y) d = c
c = py ( = y)
21
p(x, y)
.
p(y)
22
Probability
p(x, y)
= p(x)
p(y)
p(x, y) = 1/ x2 + y 2 1
= 0 x2 + y 2 > 1
p(x) =
= 0
1x2
p(x, y) dy =
1x2
1
2
dy =
1 x2
|x| 1
|x| > 1
23
= 0
|y| 1
p(y) =
|y| > 1
It is apparent that the product of p(x) and p(y) does not equal p(x, y), so the
random variables x and y are not S.I. The conditional probability is found
from Bayes theorem.
p(x|y ) =
=
p(x, y)
(1/)
{when x2 1 y 2 }
=
p(y)
(2/) 1 y 2
{when y 2 1}
1
2 1 y2
= 0
|x|
1 y2
elsewhere
It is not surprising that p(x|y) is a constant when one considers the following
interpretation.
24
Probability
.....................................
Example: Derivation of the Poisson Density
Given Events occurring at random alone a line X are governed by the
following two conditions:
In the limit X 0 the probability that one and only one event occurs
between X and X + X is given by rX, where r is a given constant
independent of X.
The probability of an event occurring in some interval X is statistically independent of events in all other portions of the line.
Problem
a) Find the probability p(n = 0; L) that no events occur in a region of
length L. Proceed by dividing L into an innite number of S.I. intervals
and calculate the joint probability that none of the intervals contains
an event.
b) Obtain the dierential equation
d
p(n; L) + rp(n; L) = rp(n 1; L)
dL
as a recursion relation governing the p(n;L).
25
1
(rL)n erL
n!
Consider dL so short that p(0) >> p(1) >> p(n > 1) in dL. But the
probabilities must sum to unity, p(0) + p(1) + p(2) + = 1, so one can nd
an approximation to p(0) which will be valid in the limit of small dL.
p(0) 1 p(1) = 1 r(dL)
The probability of an event in any sub-interval is S.I. of the events in
every other sub-interval, so
m=L/dL
p(n = 0; L) =
(1 r(dL))
m=1
ln p(n = 0; L) =
ln(1 r(dL))
m
m=L/dL
ln p(n = 0; L)
r(dL)
m=1
L
r(dL) = rL
=
dL
p(n = 0; L) = erL
26
Probability
The two intervals are S.I. so one may decompose p(n; L + L) in terms of
two mutually exclusive events.
p(n; L + L) = p(n; L)p(0; L) + p(n 1; L)p(1; L)
= p(n; L)(1 rL) + p(n 1; L)(rL)
Rearranging
p(n; L + L) p(n; L)
= rp(n 1; L) rp(n; L)
L
Passing to the limit L 0 gives
d p(n; L)
= rp(n 1; L) rp(n; L)
dL
27
c) To show that the Poisson density satises this equation, take its derivative
with respect to L and compare the result with the above expression.
p(n; L) =
1
(rL)n erL
n!
d
n
1
p(n; L) = r (rL)n1 erL r (rL)n erL
dL
n!
n!
= r
1
1
(rL)n1 erL r (rL)n erL
(n 1)!
n!
= rp(n 1; L) rp(n; L)
This solution is unique when the dierential recursion relation is supplemented by the boundary conditions
p(0; L) = erL
p(n; 0) = 0
n=
0.
.....................................
Extended Example: Jointly Gaussian Random Variables
Introduction The purpose of this example is to examine a particular joint
probability density and the information that can be extracted from it. We
will focus our attention on a physical example that might be encountered
in the laboratory. However, the origin of the eect is not of concern to us
now. We are interested instead in understanding and manipulating a given
probability density.
The System Consider an electronic circuit with all sources (power supplies
and signal inputs) turned o. If one looks at a given pair of terminals with
an oscilloscope, the voltage appears to be zero at low gain, but at high gain
there will be a uctuating random voltage that might look as follows:
28
Probability
The origin of this noise voltage is the random thermal motion of electrons
in the components. It is referred to as thermal noise or Johnson noise
and is dierent from the shot noise associated with the quantization of
charge. This noise is still present when the sources are turned on and may
complicate the detection of a weak signal. Later in the course quantitative
expressions will be derived for the amplitude of this type of noise. For the
present, observe the following features of the voltage:
1) It has zero mean.
2) Its average magnitude |v| seems relatively well dened and excursions
too far above this magnitude are unlikely.
3) The statistics do not seem to change with time.
4) There is a correlation time c such that over time intervals much less
than c the signal does not change appreciably.
5) The voltages at times separated by much more than c seem to be
statistically independent.
The noise voltage described above evolves in time and is an example of
a random process. The study of random processes is a separate eld of its
own and we will not get involved with it here. Rather, we will simply note
that by evaluating the random process at two separate times we can dene a
pair of random variables. For an important and frequently occurring class of
random processes the two variables thus dened will be described by a jointly
Gaussian (or bivariate Gaussian) probability density. It is this probability
density that we will examine here.
29
p(v1 , v2 ) =
1
v12 2v1 v2 + v22
]
exp[
2 2 (1 2 )
2 2 1 2
In the above joint probability density and are parameters. is a constant, independent of time, which governs the amplitude of the variables.
is a function of the time interval between the measurements, |t2 t1 |; it
determines how strongly the two variables are correlated and is referred to
as the correlation coecient. The magnitude of is less than or equal to
one: || 1. Physically one expects that will be close to one for very small
values of |t2 t1 | and will decrease to zero for large time separations. We
will take this joint probability density as given and examine its properties.
The variables v1 and v2 appear as a quadratic form in the exponent. Thus
lines of constant probability are ellipses in the v1 , v2 plane; when > 0 the
major axis will be along v1 = v2 and the minor axis will be along v1 = v2 ;
for < 0 the location of the major and minor axes is reversed. The ellipses
are long and narrow for ||
= 1; they become circles when = 0.
30
Probability
31
p(v1 ) =
p(v1 , v2 ) dv2
Z
1
v12
(v2 v1 )2
p
=
exp[ 2 ]
exp[ 2
] dv2
2
2 (1 2 )
2 2 1 2
|
}
{z
1
p
=
2 2 1 2
exp[
2 2 (12 )
1
2 2
exp[
v12
]
2 2
1
2 2
exp[
v22
]
2 2
The probability densities for v1 and v2 are identical in form, so one concludes that the single
time probability densities are independent of time.
Statistical Independence
p(v1 , v2 ) = p(v1 )p(v2 )
only when = 0
This implies that v1 and v2 are not statistically independent unless = 0, that is, at large
time separations between t1 and t2 .
Conditional Probability
p(v2 |v1 ) =
p(v1 , v2 )
p(v1 )
= p
1
2 2 (1 2 )
exp[
(v22 2v1 v2 + v12 )
v12 (1 2 )
2
]
2 2 (1 2 )
2 (1 2 )
1
(v2 v1 )2
p
=
]
exp[ 2
2 (1 2 )
2 2 (1 2 )
This is a Gaussian with mean v1 and variance 2 (1 2 ).
32
Probability
Compare these plots of p(v2 |v1 ) with an imaginary cut of one of the plots of
p(v1 , v2 ) by a vertical plane at constant v1 . This will allow you to picture the
33
relation
p(v2 |v1 ) p(v1 , v2 ).
The exact dependence of on |t2 t1 | depends on the details of the circuit
in which the voltage is measured.
The Correlation Function The correlation function for a random process
such as the noise voltage we are discussing is dened as
R( ) < v(t)v(t + ) > .
Here we have assumed that the statistics of the process do not change with
time so that the correlation function depends only on the time dierence,
not the actual times themselves. In our notation then = t2 t1 and
R( ) = R(t2 t1 ). We can now nd the correlation function in terms of the
parameters appearing in the joint probability density.
< v1 v 2 > =
=
=
= ( )
v1 p(v1 )
dv1
<v 2 >= 2
Thus the correlation function for the random process can be written in the
simple form R( ) = 2 ( ) and the correlation coecient can be interpreted
as the normalized correlation function for the process.
In the gures presented above has been displayed, for simplicity, as
positive. However for some random processes may become negative, or
even oscillate, as it decays toward zero. Consider the random process that
generates the following output.
34
Probability
One possible random process with these characteristics is the noise voltage
of an electronic circuit that is resonant near a single frequency. If the circuit
had a very high Q, the correlation function might oscillate many times before
falling away to zero.
.....................................
MIT OpenCourseWare
http://ocw.mit.edu
For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.