Project STT 311
Project STT 311
SCHOOL OF SCIENCES
0
STT 311 PROBABILITY DISTRIBUTION 2
(LAGOS CENTRE)
1
TABLE OF CONTENTS
1.0 Introduction
1.1 Objective
1.7 Conclusion
1.8 Summary
1.91 References.
2.0 Introduction
2.1 Objective
2
2.6 Continuous Random Variables
2.11 Conclusion
2.12 Summary
3.0 Introduction
3.1 Objectives
3.6 Moments
3
3.10 Conclusion
3.11 Summary
4.0 Introduction
4.1 Objectives
4.7 Conclusion
4.8 Summary
4
UNIT 1: PROBABILITY SPACES, MEASURE AND DISTRIBUTION
1.0 INTRODUCTION
definition and relevant working examples will be given to make the concept
1.1 OBJECTIVES.
a function that has ∆ as its domain that is a single term that gives us
notation.
5
1.3 SAMPLE SPACE AND EVENT
sample space.
Event Space: The class of all events associated with a given experiment
properties
(i). Ρ(Ω) = 1
6
(iii) If Αn, n = 1, 2 , ...................are mutually disjo int sets in Λ ,
by the event B.
Ρ (Β ) = Ρ (Β ∧ Α1 ) + Ρ (Β ∧ Α 2 )
= Ρ ( Α1 ) Ρ (Β / Α 2 ) + P( A2 ) P (B / A2 )
∧
Ρ (Β ) = ∑ i
Ρ (Β n Αi )
∧
= ∑ Ρ ( Ai ) P(B / Ai )
i
Proof:
p
( An B)
PB (A) = ≥ O;
P( B)
P(ΩnB )
PB ( Ω ) = = 1
P (B )
7
Let A1, A2……… be disjoint event in I, then
∞
[ [ U Ai[nB]
Ρ (UAi B) = 1
Ρ ( B)
Ρ U ( AinB )
=
Ρ ( B)
Lemma
∑
.
(ii) If P ( A n ) = ∞ then P (A ) = 1
1
Proof :
(i) A = ∩ U An ⊆ U Am
m =n
∞
Since = U decreaseb as n in creases, we have
m=n
∞ ∞
P( A) ≤ P U Am ≤ ∑ P ( Am )
m=n m =n
8
Since ∑P n ∠ ∞, then
∞
l im
n→∞
∑ P( Am) = 0
m =n
( ) = U I A ∞ c
∞ ∞
(iii ) P A • c
m ≤ ∑ nI= m A m
P
n = 1 m =n
∞ ∞
= ∑∏ (1 − P( Am ) = 0
n=m n=m
Hence
P(A) =1
a ≤ x ≤ b is an event in S .
Ρ ( a < x ∠ b ) defined as
The probability
b
Ρ(a <x<b) = ∫a
f ( x) d x
i. f ( X )> 0
ii ∫
R
f ( x ) dx = 1 where R = (a1b)
9
iii. f (x ) is a non - decreasing function
b
iv. ∫
a
f ( x ) dx = f (b) − f (a ) = p (a < x < b)
interval (0,1) if a point is taken in this interval say 0.45, the probability
1
P (x = 0.45) =
Uncountale po int s
1
=
No of po int s btw the int erval
1
= = 0
∝
Similarly
x 2 − x1
Pr ( x1 ∠ x ∠ x 2 ) =
Total Lenght
10
Example 1.6.1 :- Given the figure below
C
D D
Solution:
Area of ∆ CDE
Pr (Shaded portion) =
Area of ∆ ABC
( )
0 3 2 x − x 2 0 < x < 2
f (x ) = O 4
O elesewthere
, where x is the lenght of life mesasued in hours, find the expected amount of
11
Solution
2
= − ∫00 g ( x) f ( x) dn
2
= ∫0 x 2 ( 3 4 ( 2 x − x 2 ) dx
= 1.2 mg
f ( x) = ∫ O ≤ x ≤ 10
2
kxO
O elsewhere
Solution
1
If F (x) is a Pdf, ∫0
f ( x) dx = 1
∫ f ( x)dx = ∫
10 2
K x dx
0
10
Kx 3 ) (10)
3
( 0) 3
= 0
= k − k =1
3 3 3
k
= (10 3 − 00 ) = 1
3
1000 K
= 1
3
12
3
k =
1000
1.7 CONCLUSION
In this unit, you have learnt probability space, the notation of its
You also learned probability measure and its main probabilities related
1.8 SUMMARY
What you have learned in this unit are the following probability distribution
concepts.
13
EXERCISE 1.80.1 (SAE)
function.
f ( x) = ∫ 6 (x −
0
x2 ) O∠x ∠1
dscuhere
given by
2 p x =1
p = 2
x
f ( x) =
4 p x = 3
0 otherwise
14
1.9 REFERENCES / FURTHER READING
15
UNIT 2: DISTRIBUTION OF RANDOM VARIABLES SPACES
2.0 INTRODUCTION
variables into discrete and continuous random variables are high lighted,
The unit further high light Graphical representation, Joint distribution for
2.1 OBJECTIVE
with example.
16
- State the joint distributions for two random variables which are either
dependent events.
random variables.
Definition 2
Note:
17
Example : suppose that a coin is tossed twice so that the sample space
S = (HH HT TH , TT ). let X repses ent the number of heads that can come
values is called a discrete random variable. While the one which takes
arranged in some order. Suppose also that these values are assumed
p ( X = xk ) = f ( xk ) K = 1, 2, ................(1)
18
For X = xk , this reduces to equation given above while for other
values of X 1 f ( x ) = 0.
f ( x) ≥ 0
1.
2 ∑f
x
( x) = 1
Where the sum in equation (2) is taken over all possible values of x
f ( x) = P ( X ≤ x )
−∞ ∠x ∠ ∞
2. l im F ( x) = 0; lim f ( x) = 1
x→∞ η →∞
3. f ( X ) is continous fromthe right ie lim f ( x + h) for al x .
ης∞
19
2.5 DISTRIBUTION FUNCTION FOR DESECRATE RANDOM
VARIABLES
x in (−, ∞, ∞)
p ( x) = P ( X ≤ x ) = Σ f ( x)
v ≤η
Where the sum is taken over all values µ taken on by X for which
0 − ∞ < x < x1
f (x ) x1 ≤ x ≤ x2
1
F ( x)
[ f ( x1 ) + f ( x2 ) x2 ≤ xη ∠ x2
[ F ( x ) + − + F x ) x ≤ x ∠ ∞
1 n n n
random variable x when a coin is tossed twice; assuming that the coin
is fair.
Solution
p ( HH ) = 1 , P ( HT ) = 1 , P (TH ) = 1
4 4 4
and P(TT ) = 1
4
20
Then, P ( X = 0) = P(TT ) = 1
4
P ( X = 1 = P ( HT ) + P (TH )
= 1 + 1 = 1
4 4 2
P ( X = 2 ) = P ( HH ) = 1
4
x 0 1 2
f (x) 1 1 1
4 2 4
Example 2.5.2 = (a) find the distribution for the random variable X
Solution:
21
0 − ∞ ∠ x ∠ 0 sin ce x1 = 0 from the above table
1 = 1
4 0 ≤ x ∠ 1 sin ce f ( x1 ) 4
f (η ) 3 1 ≤ x ∠ 2 sience f ( x1 ) + f ( x2 ) = 3
4 4
ie 1 + 1 = 3
4 2 4
1 2 ≤ x ∠∞ sin ce 1 + 1 + 1 = 1
2 4 4
b.
f (x)
3
4
½
1
2
1 31
4 4
0
1 2 x
noted.
22
2. Because of the appearance of the graph it is often called a staircase
1. f ( x) ≥ 0
∞
2 ∫f
∞
( x)dx = 1
then, the probability that X takes on only one particular value is zero,
23
b
p (a ∠ x ∠ b ) = ∫ f ( x) dx
a
Example 2.6.1 (a) find the constant C such that the function
cx 2 0 ∠ x ∠ 3
f (x ) =
0 otherwise
b. Computer P (1 ∠ x ∠ 2 )
Solution
C = 1/9
2
b. P (1∠ x ∠ 2) = ∫
1
1 x 2 dx
9
24
x3 2 ( 2 ) − (1)3 3
27 ∫1
= =
27 27
8 1 7
= − =
27 27 27
Since f ( x) ≥ O, the curve cannot fall below the x-anis the entire area
bonded by the curve and the x-anis must be I because of the second
∞
property i.e ∫
∞
f ( x)dx = 1) .
Geometrically the probability that x is between a and b, i.e p (a < x < b),
is the represented by the area shown shaded from the first figure below
25
The destitution function F ( X ) = p ( X ≤ x) is a monotonically increasing
26
2.8 JOINT DISTRIBUTIONS
We shall consider the typical case of two random variables that are either
Discrete case: - If x and Y are two discrete random variables, we define the
P ( X = x, Y = y ) = f ( x, y )
Where (1) f ( x, γ ) ≥ O
(2) ∑∑ f
x y
( x, y ) = 1
Suppose that x can assume any one of m values x1, x2 , − − xm and y can
27
X
Y Y1 Y2 --------- Yx Total
2
P ( x = η j ) = f1 ( x j ) = ∑fk −1
( xj y k )
entrance right hand column or margin from the table above similarly
the bottom row or margin of the probability table from the two
28
which are obtained from the margin of the table are refer to as the
∑f
j =1
1 ( xi ) = 1 ∑f
k =1
2 ( yk ) = 1
m η
∑ ∑f
j =1 K =1
( xj , yk ) = 1
This is simply the statement that the total probability of all entries is
probability table.
f ( x1y ) = P ( x ≤ x, y ≤ y ) = ∑ ∑
µ ≤ η ν ≤ y
f ( x1 ν )
and yk ≤ y.
CONTINUOUS CASE:
The case where both variables are continuous is obtained easily by analogy
with the discrete case on replacing sums by integrals thus the joint
defined by
(1) f ( x, y ) ≥ O
29
∞ ∞
(2) ∫ ∫
−∞ −∞
P ( x1 y ) dx dy = 1
30
More generally, if A represents any event, there will be a region RA of
the xy plan that corresponds to it. In such case we can find the
P (A) = ∫ ∫
RA
f ( x1 y ) dxdy )
F ( x, y ) = p ( X ≤ x, Y ≤ y ) = ∫
µ = −∞
∫
v = −∞
f (u, v) du dv
d F ( x)
= f ( x) ,
dx
∂2 F
= f (x,y)
∂ x dy
∞ y
P (y ≤ y ) = f 2 ( y ) = ∫
µ = −∞
∫
v = −∞
f (u , v) dudv
The two equation above are called the marginal distribution functions or
31
The derivative of the equations with aspect to x and y are then called the
Suppose that x and y are discrete random variables. If the events X = x and
Y = y are independent events for all x and y, then we say that x and y are
P (X = x, Y = γ ) = p ( X = x) p (Y = y )
Or equivalently
f( x, y ) = f1 ( x) f 2 ( y )
(which are then the marginal probability function of X and Y) X and Y are
dependent. If X and Y are continuous random variables, we say that they are
32
P (X ≤ x, Y ≤ y ) = p ( X ≤ x ) p ( Y ≤ y ) or equivalently
F ( x1 y ) = F1 ( x) F2 ( y )
continuous independent random variables, it is also true that the joint density
of y alone, f2 (y) , and these are the (marginal) density functions of x and y,
respectively.
AnB)
P(B/A) = P (
P( A)
If x and y are discrete random variables and we have the events (A: x = x)
33
P(Y = y X = x ) =
f ( x, y )
f1 ( x )
f ( x1 y )
F(y x) =
f 1 ( x)
f ( x1 y )
F ( x y) =
f 2 ( y)
f ( x 1 y)
f ( ylη ) =
f1 ( x)
34
Example 10.1
C
f ( x) = , where − ∞ < x < ∞ .
( x + 1)
2
Solution
∞
c. We must have ∫
−∞
f ( x) dx = 1 ie
∞ c
∫ dx = C tan − 1x ∞
−∞ x + 1 2 −∞
∏ ∏
= C − − = 1
2 2
=C∏= 1
C = 1
∏
1 3
b. If ≤ x 2 ≤ 1 , then either ≤ x ≤ or
3 3
3
−1≤ x ≤ -
3
35
1 − 53 dx 1 1 dx 2 1 dx
∫ + ∫ = ∫
3
∏ −1 x 2 +1 ∏ 53
3 x 2 +1 ∏ 53
3 x 2 +1
2 −1 −1 J3
= tan ( 1) − tan ( )
∏ 3
2 ∏ ∏ 1
= − =
∏ 4 6 6
Example 2:10:2
Solution :
u 1 µ du
f ( x) = ∫
∞
f (u) du =
∏ ∫∞ u 2 + 1
−∞
1 −1
x
=
∏
tan u ∫
∞
=
1
∏
tan [ −1
x − tan −1
(−∞) ]
1 −1 ∏ 1 1 −1
= tan x+ = + tan x
∏
2 2 ∏
Example 2:10:3
1 −2µ
f ( x) = −e x ≥ 0
0 η 0
36
Find (a) the density function (b) the probability that x > 2 and (c)
Solution :
f ( x) = 1 − e − 2 x when x ≥ 0
a. d − 2x
( f ( x) = 0 − (−2 ) e − 2 x = 2e
dx
d 2e − 2 x x ≥ 0
2 > f ( x) = f ( x) =
dx 0 x< 0
∞
− 2η − 2µ ∞
(b) p (x > 2 ) = ∫ 2
2e dx = − e
2
. = − e − 2 (∞ )
−4
= − eo
+ e = e− 4
4
− 2µ
4 o
o dx + ∫ 2e dx
p (−3 < x ≤ 4) = ∫ f (x) dx = ∫
o
c. −3 −3
= − e − 2x 4
= 1 − e −8
o
OR p (−3 < x ≤ 4 ) = P ( x ≤ 4 ) − P x ≤ − 3)
P (4) − P (−3)
= (1 − e − 8 ) − (o)
− 8
= 1− e
o ≤ x ≤ 2, o ≤ y ≤ 3 , and f ( x, y ) = o othewise
37
b. Find p (x = 2, y = 1). (c) find p (x >1, y < 2
Solution:
The sample points x, y ) for which probabilities are different from zero
y 0 1 2 3 Totals
x
0 0 C 2c 3c 6c
1 2c 3c 4c 5c 14c
2 4c 5c 6c 7c 22c
C=1
42
F ( x1 y )
b. P (x = 2 , Y = 1) = C (2 x 2) + 1) +
grand total
5c
5c +
42c
5
= 5c +
42
38
c. p (x ≥ 1 , y ≤ 2 ) = ∑ f ∑ f ( x, y )
x ≥=1 y≤2
Solution
f ( x, y ) (2 x + y ) / 42
f ( y x) = =
f1 ( x) f1 ( x)
So that with x = 2
(4 + y ) / 42 4 + y x 21
f (y 2 ) = =
11 11 x 42
21
4+ y
=
22
4 +1
p ( y = 1 x = 2) = f (1 2) =
(b) 22
= 5
22
39
2.11 CONCLUSION
functions for discrete and continuous random variables. You also learned
random variables.
2:12 SUMMARY
In this unit distribution of random variables spaces that you studied included
Exercise 2.12 .1 (S A E)
40
Suppose that a pair of fair dice are to be tossed, and let the random
distribution for x.
variables x and y is
{C x y o ∠ x ∠ 4, 1 ∠ y ∠ 5
f ( x, y ) =
0 otherwise
(a) Find P ( 1 ∠ X ∠ 2, 2 ∠ Y ∠ 3)
41
UNIT 3 EXPECTATION OF RANDOM VARIABLES.
3.0 INTRODUCTION.
Expected value for discrete and continuous random variables are stated.
Moment and Moment generating functions for random variables are also
also learned and relevant working examples on each concept are given to
3.1 OBJECTIVE
42
4. State Variance and Standard Deviation for Discrete and Continuous
Random Variables.
variables.
= ∑x f (x) (1)
distribution, then
expectation of x is defined as
∞
E (x) = ∫−∞
x f ( x ) d x − − − ( 2)
43
Provided that the integral converges absolutely. Where f (.x) is the value
1 1 1 1 1 1
expectation is E ( x) = 1 x + 2x + 3x + 4x + 5 x + 6x
6 6 6 6 6 6
1 2 3 4 5 6
= + + + + +
6 6 6 6 6 6
21
= = 3 1 (3.5)
6 2
4 1 2
E ( x) = ∫ x. 1 dx = =3
1
x
2 2 4
2
For continuous case using the above example find the expectation of
44
Solution
E [g ( x) ] = ∑ x 2 f ( x)
f ( x) = 1 2, 3, 4, 5, 6
(
f ( x) = 1
6
∴ E [g ( x ) ] = 1 x
1 1 1 1 1
+ 4 x + 9 x + 16 x + 25 x + 36 x 1
6 6 6 6 6 6
= 15 16
Similarly for the random variable with the uniform density function
E [g ( x)] = E ( x 2 ) =
4
∫
2
x 2 f ( x) dx
4 1
= ∫
2
x 2 ( ) dx
2
4 1 3
∫ 12 x dx =
2 4
= x 2
2 6
= 9 13
E (X + Y) = E (x) + E (Y)
45
Theorem 3.3.3: If x and y are independent random variables, then
E (X Y) = E (X) E (Y)
THEOREM 3.3.4
E [(ax + b ) ]= ∑
η
η
(in ) aη −ι bι E ( xη −ι )
ι=o
For instance, if n = 1,
E [(an + b) n ] =
n
∑ (1ι ) a
ι =o
1−ι
bι E ( x1−ι )
()
= 1 a E (x ) + 1 b E (1)
o
() 1
= a E (x) +b
If I= 2
[ ] ∑ (2) a b E [x − ]
2
2 −ι
E (ax + b ) 2 = i 2 i
ι
ι =o
( )a E (X )+ ( )ab E ( X ) + ( ) b E (1)
20
2 2 2
1
2
2
2
E [g ( x, y )] = ∑ ∑ g ( x, y ) f ( x , y ) − − ( 4)
η y
Or E [g ( x, y )] =
∞ ∞
∫ ∫ −∞ −∞
g ( x, y ) f ( x, y ) dx dy − −(5)
46
In equation (4) f ( x, y ) is the value of joint probability function of x and
probability density
σx = Var ( x ) = E ([X ])
− x ) 2 − .......... ....................... (7)
In the special case of (8) where the probabilities are all equal, we
47
Which is the variance for a set of n numbers x1...................... xn
Theorems 3.5.1
σ 2 = E ( x − µ ) 2 x) = E ( x2 ) − µ 2 =
= E ( x 2 ) −( E ( x) 2
where µ = E ( x) .
48
3.6 MOMENTS
The rth moment of a random variable X about the mean µ , also called the
µr = ∑ (x − µ ) r
f ( x) (discrete) .................................................................(15)
var iabl ; e
∞
µr = ∫ ( x − µ ) r f ( x) dx (continuous) .................... ..........................(16)
−∞ var iable
The rth moment of x about the origin, also called the rth raw moment, is
defined as
The zero moment and the first moment about the mean are respectively 1
And µ 1 = E [ ( x − µ ) 1 ] = E ( x) = E ( µ ) = µ − µ = o
[
µ 2 = E (x − µ )2 = σ 2 ]
This indicate the strength on dispersion of the distribution generated.
49
Generally moment about the mean describe the shape of the distribution of a
random variable.
M x (t ) = E (e tx
) = ∑e t x
f ( x)
( discrete var iable )
.......... ................................(18)
∞
M x (t ) = E (e tx ) = ∫
t x
e f ( x) d x .................... ..............(19)
−∞ ( continuous var iable )
tx t 2 x 2 t 3 x3 tr x r
but e txη = 1 + + + + −−− −−−
! 2! 3! r!
t 2 x2 tr x r
This for the discrete case Μx (t ) = ∑ 1 + tx + + − − −+ − − f ( x)
21 r1
2 r
t t
⇒ Μ x (t ) = ∑ f ( x + t ∑ x f ( x) + ∑ x 2 f ( x) + − − − ∑ x r f ( x) + − − −
2! r!
t2 tr
Μ x (t ) = 1 + µ 1t + µ 1 + −−− µ1 + .................... ........................................ . (20)
2 2! r!
50
You observed that the maclaurins series of a function Μx (t ) with coefficient
tr
is the rth derivation of the function with respect to t at t = o.
r!
variation
dr Μx (t )
µ r1 =
dt r t = o
x+a
function of (t ) = t
at
is M x+a
e b
Mx
b
b
b
Μ x + y (t ) = M x (t ) M y (t ) ..............................................................(21)
51
3.9 CHARACTERISTIC FUNCTIONS
generating function because it can be need to prove both the weak law of
large numbers and the Central limit theorem which will be treated in the next
unit.
ψ (t ) =
tx
∫ ei d F ( x) = ∫e
itx
p (dx)
R
= ( )
E eitx = E eitx ( )
Where eitx = cos t + i sin t x
( )
E eitx = E (cos t x ) + i E sin t x
(ii) Ψ (o) = 1
Since e itx = 1
52
Proof:
(i) φ (t + h) − φ (t ) = ∫
−∞
∞
{e i (t + h)
x ei
th
}p (d x)
∞ −h
≤ ∫
−∞
e i (t x − e itx p (dx)
∞ ∞
= ∫− ∞ e itx e iηx − p (dx) ≤ ∫− ∞ e i ηx p (dx)
h → 0 ∞∫
e itx e − 1 p (dx) = ∫ lim e − p (dx) = 0
−∞ h →ο
Note:
(i) e i η x − 1 ≤2
Thus, φ (t + h) − φ (t ) → o independent of t.
(iii) e itx
= cos t x + ι sin t x = cos 2 t x + sin 2 t x = 1
53
Proof :
random variables.
Then
ψ sn (t ) = ψ x (t ) n
Is finite for all variable x and all real number t. The reason being that eit is
formula:,
η
If X is integer valued random variable then f x (n ) =
1
∫ e itn
ψ x (t ) dt
2∏ ∏
1 ∞
If x is a continuous random variable, then f x ( x) = ∫ eit ψ x (t ) dt
2∏ ∞
∞
assuming ∫ ψ x (t ) ∠ ∞
∞
54
e. Properties of characteristic function enable us to prove both the
Properties (c) (d) and (e) are important advantages of functions over
(f) If two random variables have the same characteristic function they
µ x ( (t ) = d n
n) E (e itx
dt n
n
{
) E (ix ) eitx }
µ x ( n ) (0 )
Thus E (x n ) =
in
Solution
∞
− βx
ψ x (t ) =E (e i ) = βe
tx
∫
itx
e dx
∞
∞
( β − it ) x β
= β ∫ e− dx =
∞ β − it
characteristics function of X.
55
Solution
(t ) ) = E = ∫−1e 1 1
ψ x dx itx
2
1 e i tx 1
= t≠0
2 it −1
1 ei − e − it
t
sin t
= =
2 it t
e it
= cos t + i sin t ,
Note: e− it
= cos t − i sin t ,
− it
e it
−e 20 sin t.
1 1x1 ≤ a
f ( x = 2a
0 otherwise
Solution
∞ 1 a
E (eitx ) = ∫ f ( x) dx = ∫
itx itx
e e dx
∞ 2a a
56
Solution
∞ x
1 1 1 1
We have E ( x) = ∑ x = + 2 +3 +
x = 1 2 2 4 8
1 1 1 1
Then S = + 2 + 3 + ……………………………….
2 4 8 16
1 1 1 1 1
By subtracting S = + + + + ...................... = 1
2 2 4 8 16
Therefore, S = 2
2e − 2 x x> 0
Density given by f (n) =
0 x≤0
∞ ∞
(∞) E ( x ) = ∫ x f ( x) dx = ∫ x ( 2e − 2x
) dx
∞ ∞
∞
− 2x
= 2 ∫∞
xe dx
e
= 2 ∫ ( x) − 2 x − (1)
e− 3 x ( ) ∞ 1
2 4 ∫∫η =
2
∞ ∞
−2
E( x2 ) = ∫ −∞
f ( x) dx = 2 ∫ο x2 e xdx
(b) e− 2 x −2x
e− 2µ
= 2( x 2 )
− 2
( )
x e
− 2
4
+ 4 ∫∫
e
=
1
8 0 2
57
Example 3.9.6 Find (a) the variance, (b) The standard deviation of the
Solution
91 7 91 49
Var ( X ) = Vor (Y ) = − 2
= −
Then 6 2 6 4
= 35
12.
var ( x + y ) =
35
ie σ x1+ y
=
6
Example 3.9.7: The random variable x can assume the values 1 and -1
58
Solution:
(a.) ( )= e
E e it
x t (1) 1
+e
2
t ( −1) 1
2
et + e −1
1
=
− .............................. ......................(1)
2
t2 t3 t4
(b) we have − et = 1 + t +
+ + .......... .................... .....( 2)
2! 3! 4!
t2 t3 t4
But from (2) M x (t ) = 1 + µ 1 2 + µ 1 + µ 13 + µ1 4 + .................... ..................
2! 3! 4!
Then compering (1) and (2) we have
µ = 0, µ 1 2 = 1, µ 1 3 = 0, µ 1 4 = 1
The odd moment are all zero, and the even moments are all one
− 2x
2e x≥0
f (x) ) =
0 x<0
Solution:
( ) ∫ ∞ tx
(a ) M x (t ) = E etx = e f ( x) dx
−∞
∞
−2x
= ∫
tx
e ( 2e ) dx
∞
∞
= 2 ∫ e (t − 2 ) x
dx
∞
(t − 2) x ∞
= 2e
t−2 ο
59
2
= assu min g t < 2
2−t
2 1 t t2 t3 t4
= =1+ + + +
2−t 1− t 2 4 8 16
2
t t3 t4
But Mx (t ) = 1 + µt + µ 12 + µ 13 + µ 14 + .......... ........................
2! 3! 4!
1 1 3 3
µ21 = , µ 3 = , µ 14 =
2 4 2
3.10 CONCLUSION
for discrete and continuous random variables. You also learned Variance
are stated.
fully treated and related working example on each concept are easily
60
3.11 SUMMARY
variables
variables.
Random Variables
Variables
1
x 0 <x< 2
f ( x) = 2
0 otherwise
61
3.1.2 TUTOR MARKED ASSIGNMENT (TMA)
E (3 x 2 − 2 x)
Materials.
62
LIMIT THEOREM
4.0 INTRODUCTION
The purpose of this unit is to acquaint the students with the liquid
4.1 OBJECTIVES
continuous)
63
Having mean µ and variance σ 2 , which are finite. Then if epsilon (e)
is a positive number,
σ2
P (1 x − µ ≥ E) ≥
ε2
or with ε = k σ
1
P(x−µ 1 ≥ kσ ) ≤
k2
Proof :
∞
σ 2 = E ( X − µ ) 2 = ∫ (x − µ)
2
f ( x)dx
∞
Since the integrand is nonnegative, the value of the integral can only
Therefore,
σ2 ≥ ∫ ) ≥E ( x − µ ) 2 f ( x) dx ≥ ∫ ε 2 f ( x) dx
1x − µ
1− µ 1
2E
= ε2 ∫f ( x) dx
1x − µ ) ≥ E
Hence,
64
σ2
P (1 x − µ 1 ≥ ε) ≤
ε2
Bi (n , p ) , then
random variable Bi (1, p ), Bi (2, p ), Bi (3, p ) .................... ...... → X
Bi (n, p ) tends to normal
()
Fn = Pr ( X ≤ x) = ∑ n p x q n
n
−x
→ F
If Xn
→
l
X ie Fn ( X ) → F distribution function of X Then
(1) Cn (t ) → C (t )
variable Xn
∫ gd Fn ( x) → ∫ gd F
Definition 2: The sequence {Xn} of random variables is said to
65
P { Xn − X >ε } → 0 as n → ∞ for an ε > 0 . we indicate this by Xn
→
P
X
Def int ion 4 : The sequence {Xn} of random var iable is said to converge with probablity
one(or almost surely , a.s.) to a cons tan t C if {
P lim X
n→∞
n =c } =1
Note:
P { Xn −x ≥ ε }= P {− 2 x ≥ ε }= ε
P x ≥ → 0 as n → ∞
2
Hence Xn
→p
X
(ii) If X n →
P
X then Fn → F . letting ε → 0,
We see that Fn ( x) → F ( x)
66
4.4 DEMOVRE’S THEOREM
Xn − np
Yn = → N (ο ,1) as n → ∞
npq
Pr oof Pr ( X n = r ) = n P r q n − x ()
r
Cn (t ) xn = E (ei tx )
∞
= ∑e
x=0
itx
Pr( Xn = x)
e ( )P
∞ itx
= ∑
µ =0
n
r
x
q n−x
( ) (e )
∞
∑
x
= n
r
itx
P q n− x
µ=0
= Pe ( it
+q ) n
t yn
Therefore characteristic function of Yn = E (ei )
itX n itnp
= E e −
jnpq npq
np
itxn
= e − it q
E e
npq
n
−it
np
it
+ q
i.e Cy n (t ) = e q p e vnpq
67
it
n log Pe
np
→ log C yn (t ) = − it +
evnpq
+1 − p
q
np it
= − it +n log 1 + p e 1
q jnpq −
z2
But ez = + z + + .......... .............................. .................... ....
2!
z2 z3
And log (1 + z ) = z − + .................... .............................. ..............
2! 3!
∴ log y n (t ) = it
np it
+ n log 1 + p +
(it )2 + (it )3 + }} 0 ⇒ order
q npq
2npq npq)3!
np it t2 I p2 −t2 1
= − it + n Log 1 + P − + 0 3 .... − npq + 0 2
q npq 2npq n 2 2 n
np nt 2 p np 2 t 2 1
+ O
np
= − it + it − +
q q 2npq 2npq n1
2
t2 1
= − + 0
2 n
t2
−
Cy (t ) → e 2
as n → ∞
t2
−
but e 2
is cheracteristic function of a stadard normal distrubtion N (0.1)
∴ − Yn →
L
N (0,1)
Theorem
68
np = λ , Xn →
L
Poison (λ )
Re call : Cxn (t ) = Pe it + q ( )
n
n
λ it
e + + 1 − λ = 1 + λ e it − 1
n
n
( )
n n
n
z
lim 1 + = e z
n
λ
n
lim Cxn (t ) = lim 1 + e ( it
−1 )
n →∞ n →∞ n
P →0
np →λ
= e λ eit − λ
X 1 + X 2 + ............................. + xn ) − µ
Let Yn = (
n ∂
x − µ n
=
∂
→ N (0,1)
L
Then yn
X 1 + X 2 + .............................. + Xn
X =
n
69
n X = X 1 + X 2 + .......... ........................ + X n
nX − µ
nX =
nσ
Poof:
( )
⇒ E (e itx ) = E e itx 2 = .............................. ....... =E e itx = C (t )
1
( ) n
Ctn (t ) = E (e )
it y
n
X + .......... ..... X n − µ
= E e Xi + 2 .
σ n
itu itx1
itx 2 itxn
− n σ
= e E e . E e − − − − − E e
nσ n σ n σ
n
itu t
C− (t ) = e −
C
n σ n σ
Yn
(it ) 2
But C (t ) = 1 + α1 it + α 2 .......... .................... .....................
2!
Where d1 = µ 11 = µ
σ 2 = µ2 = µ 2 + σ 2
it
2
itu tu
µ 21
∴C (t ) = e − 1 + i + + 0 (t 3 )
Yn
nσ nσ nσ 2!
70
Taking log to get e
n
itµ itµ t 2 µ12
log e CY n (t ) = −
+ n log 1 + − − + 0 (t 3
)
nσ nσ 2σ n 2
X2 X3
But log (1+X) = X – + − ................
2 3
ctµ
t 2µ1
( itµ ) 3
∴ log CY n (t ) = − it
n
µ +n∫ nσ
− ( +− + ) ........
σ 2 nσ 2σ 2 n
t2
=- ( µ 2 2 − µ 2 ) as n → ∞
2σ2
= − t2
t2
∴C Yn
(t ) → e − (remove the log )
2
Hence Y n →
L
N (0,1)
Xn
→
P
C
probability to X if Xn → X →
P
0
71
4.6 KHINCHINE’S THEOREM
x1 + X 2 + X 3 + .................... ................ + Xn
Let X n =
n
Then X n →
P
µ
Poof:
C (t ) = E (e it x
1 )
= 1 + i t µ + 0 (t 2 )
Let C X
(t ) = E e ( it x
n )
= E e ( + X it ( x1
2 + .................... + X n )
= E (e ) E(e it x
1
1tx
2 ). E (e itx 2 )
n
=C
t
= E (e it x1 ) .E (e itx2 ). E e itx n ( )
n
itµ
n
= 1 +
+ 0 t2 ( )
n
= y C xn (t ) → eitµ as n → ∞
hence, X n →
P
µ
meaning l im P
n →∞
[ x −µ >k =0 ]
eitµ is the characteristics function of a random variables taking the
72
THEOREM 4
∞
612
E Xn) = ), Var ( X o) = 01 − 2 and ∑1 12 < ∞,
n
1
Then the sequence
n
∑ xi
e =1
converge to 0 almost easily ie)
Proof:
Xn
Let Yn = n = 1, 2, .............................
n
(Yn) = ∑
σ 2
n
E (Yn) = 0 and ∑Var 2
< ∞,
n
∞ ∞
Xn
∑ Yn
n =1
= ∑
n =1 n
1
lim
n →∞ n
∑ Xi = 0 a .e
Corollary 1
Sn ∑ Xi
= 1
→ο almost surely
n n
73
Corollary Π
Sn
Then → µ almost surely
n
Example 4.61
1
0 with Pr ob n
Suppose X
η=
1
1 with prob 1 −
n
X = 1 with probability 1.
Solution.
1
0 , with prob − n
Xn − X =
1 with prob 1
n
0 X < 0
1
P { X n− X ≤ X } = 1 −
0≤ X < 1
n
1 X ≥1
74
Xn − X ≤ X 0 x < 0
Lim P =
1 x ≥ = 0
X n →
P
X
0, x < 0
1 0, X < 1
Fn ( X ) = , 0 ≤ x < 1, F ( x) =
n. 1, X ≥ 1
1, X ≥ 1
F n ( x) → F ( x) for all X
Note: X n →
d
Xn → X
Example 4.6.2
− − λ
E X n − λ 2
} = Var ( X n ) = → o as n → ∞
n
− −
Therefore, X n → λ or X n
Means quare
→ λ
quadratic mean
Example 4.
−
p X −λ >l }≥ Var ( X n )
l
λ
= → 0 as n → ∞
nl 2
75
Therefore,
X n →
P
λ
EXAMPLE 4.6.3
By chebychev’s inequality
( Xn − X 2
)
P X n − X >l} ≤ E → 0 as n → ∞
l2 ∞
Thus X n →
P
X
4.7 CONCLUSION
In this unit you have learned chebyshev’s inequality, the proving and the
Moreover, the weak and strong law of large numbers are discussed the and
76
4.80 SUMMARY
2. Chebyshev’s Inequality
4. Demovre’s Theorem
EXERCISE: 4.80.1 (S A E)
(a) P( x- 3 ≥ 2 ), (b) P ( x -3 ≥ ).
EXERCISE 4.80.2
Show that the (weak) law of large numbers can be stated as.
Sn
Lim P ( − µ <l) =
n
77
4.90 TUTOR MARKED ASSIGNMENT (TMA)
converge
Hill.
78