C0 English
C0 English
Review of Probability
and Statistics and Introduction to
Econometrics
(Course: Econometrics)
Lê Phương
1 Probability Theory
Random Variables and Probability Distributions
Characteristics of Random Variables
Common Probability Distributions
Random Vector
2 Applied Statistics
Characteristic Statistical Parameters
Parameter Estimation Problems
Hypothesis Testing Problems
3 Introduction to Econometrics
Introduction to Econometrics
Methodology of Econometrics
Research Data
Regression Relationships
RVs and Probability Distributions
X x1 x2 ··· xn ···
P f (x1 ) f (x2 ) · · · f (xn ) · · ·
Properties
1 0 ≤ f (xi ) ≤ 1 for xi ∈ X (Ω),
P
2 f (xi ) = 1,
i
P
3 P(a < X ≤ b) = f (xi ).
a<xi ≤b
RVs and Probability Distributions
F (x) = P(X ≤ x) ∀x ∈ R.
Interpretation
The cumulative distribution function gives the percentage of the
values of X that are less than or equal to the real number x.
Calculation
X
f (xi ) if X is discrete,
xi ≤x
F (x) = Rx
f (s)ds if X is continuous.
−∞
RVs and Probability Distributions
Definition
The expectation (or mean) of a random variable X , denoted by E(X )
or EX , is defined as:
X
xi f (xi ) if X is discrete,
i
E(X ) = +∞R
xf (x)dx if X is continuous.
−∞
Meaning
The expectation represents the average value of the random variable
X in the experiment.
Characteristics of Random Variables
Properties of Expectation
1 E(a) = a for any constant a,
2 E(a + bX ) = a + bE(X ),
3 If g is a function,
P
g(xi )f (xi ) if X is discrete,
i
E(g(X )) = +∞
R
g(x)f (x)dx if X is continuous.
−∞
Meaning
The variance represents the degree of dispersion of the random
variable around the mean value.
Characteristics of Random Variables
Properties of Variance
1 V (X ) = E(X 2 ) − (EX )2 ,
2 V (a) = 0 for any constant a,
3 V (a + bX ) = b 2 V (X ),
4 If X , Y are independent, then:
V (X ± Y ) = V (X ) + V (Y ).
Common Probability Distributions
Normal Distribution
A continuous random variable X is said to have a normal distribution
with parameters µ and σ 2 (µ ∈ R, σ > 0), denoted by X ∼ N(µ, σ 2 ), if
its probability density function is
1 (x−µ)2
f (x) = √ e− 2σ2 .
σ 2π
X is called a standard normal distribution if X ∼ N(0, 1).
Common Probability Distributions
Chi-Squared Distribution
A random variable X has a chi-squared distribution with n degrees of
freedom, denoted by X ∼ χ2 (n), if its probability density function is
n 1 e− x2 x n2 −1 , for x > 0;
f (x) = 2 2 Γ( n2 )
0, for x ≤ 0.
R∞
where Γ(x) = 0
t x−1 e−t dt is called the Gamma function.
Common Probability Distributions
Student’s t-Distribution
X is said to have a Student’s t-distribution with n degrees of freedom,
denoted by X ∼ t(n), if its probability density function is
Γ( n+1
− n+1
2 )
√1 x2 2
πn Γ( n2 )
1+ n , for x > 0;
f (x) =
0, for x ≤ 0.
Fisher’s F-Distribution
X is said to have a Fisher’s F-distribution with n and m degrees of
freedom, denoted by X ∼ F (n, m), if its probability density function is
n m
n 2 m 2 Γ( n+m 2 )
n−2 n+m
2 (m + nx)− 2 ,
f (x) = Γ( n2 )Γ( m2 ) x for x > 0;
0, for x ≤ 0.
Definition
A mapping X = (X1 , X2 , . . . , Xn ) : Ω → Rn is called an n-dimensional
random vector.
Classification
An n-dimensional random vector X = (X1 , X2 , . . . , Xn ) is called
continuous or discrete if all component random variables
X1 , X2 , . . . , Xn are continuous or discrete.
Random Vector
Discrete Joint Probability Distribution
The joint probability distribution of a two-dimensional discrete random
vector (X , Y ) shows X (Ω) = {x1 , x2 , . . . }, Y (Ω) = {y1 , y2 , . . . }, and
the joint discrete probability function is
f (xi , yj ) = P(X = xi ; Y = yj )
Properties
1 0 ≤ f (xi , yj ) ≤ 1 for xi ∈ X (Ω), yj ∈ Y (Ω),
P
2 f (xi , yj ) = 1,
i,j
P P
3 P(a < X ≤ b; c < Y ≤ d) = f (xi , yj ).
a<xi ≤b c<yj ≤d
Random Vector
Calculation Formula:
X X
f (xi , yj ) if (X , Y ) are discrete,
xi ≤x yj ≤y
F (x) = Rx Ry
f (t, s)dsdt if (X , Y ) are continuous.
−∞ −∞
Random Vector
Theorem
Two random variables X and Y are independent if
Covariance
The covariance of two random variables X and Y is
Note:
1 If X , Y are independent, then Cov(X , Y ) = 0 (However, the
reverse is not true).
2 Cov(X , X ) = Var (X ).
Interpretation: Covariance measures the extent to which X and Y
move together or in opposite directions.
1 Cov(X , Y ) > 0: X and Y move together.
2 Cov(X , Y ) < 0: X and Y move in opposite directions.
Random Vector
Correlation Coefficient
The correlation coefficient of two random variables X and Y is
Cov(X , Y ) E(XY ) − E(X )E(Y )
ρ(X , Y ) = p = .
Var (X )Var (Y ) σ(X )σ(Y )
Covariance Matrix
The covariance matrix (also known as the correlation matrix) of the
random vector (X , Y ) is the matrix
" #
Cov(X , X ) Cov(X , Y )
Var (X , Y ) =
Cov(Y , X ) Cov(Y , Y )
" #
Var (X ) Cov(X , Y )
= .
Cov(Y , X ) Var (Y )
Unbiased Estimation
The statistic θ̂ is called an unbiased estimate of θ if E θ̂ = θ.
Efficient Estimation
The statistic θ̂ is called an efficient estimate of θ if it is an unbiased
estimate and has the smallest variance among all unbiased estimates
of θ.
Consistent Estimation
P
The statistic θ̂ is called a consistent estimate of θ if θ̂(X1 , ..., Xn ) −→ θ.
Thus, for large n, with probability close to 1, we have: θ̂ ≃ θ.
Parameter Estimation Problems
With a specific sample (x1 , x2 , ..., xn ), θ̂1 takes the value θ1 and θ̂2
takes the value θ2 . The interval (θ1 , θ2 ) is called the interval estimate
of θ, where
• 1 − α: confidence level of the estimate,
• (θ1 , θ2 ): confidence interval of the estimate,
• ε = θ2 −θ2 : precision (error) of the estimate.
1
XXX
XXXReality H0 True H0 False
Decision XXX
X
Do not reject H0 Correct decision, Type II error,
probability 1 − α probability β
Reject H0 Type I error, Correct decision,
probability α probability 1 − β
In hypothesis testing, if the probability of Type I error decreases, the
probability of Type II error increases. The significance level, denoted
by α, is often set before testing to control the probability of Type I
error.
Introduction to Econometrics
Econometrics
• Estimating and measuring economic relationships.
• Comparing economic theories with real-world data to test the
validity of economic theories.
• Forecasting economic variables.
5. Testing:
• Using statistical testing methods: test the values of parameters, the
nature of relationships,
• Test the accuracy of the model,
• If not appropriate: return to previous steps. Modify or construct a
new model to achieve the best results.
6. Using the model:
• Based on the results considered satisfactory: forecast
relationships, and objects under defined conditions.
• Evaluate decisions.
Research Data
Types of Data
• Time series data: data of an economic variable at multiple points
in time.
• Cross-sectional data: data of multiple economic variables at the
same point in time.
• Panel data: a combination of the two types above.
Data Sources
• Experimental data
• Non-experimental data
Regression Relationships
Regression Relationship
Regression studies the dependence of one economic variable
(dependent variable) on one or more other economic variables
(independent variables, explanatory variables) based on the idea of
estimating the mean value of the dependent variable given the
known values of the independent variables.
Thus:
• Independent variables have predetermined values
• The dependent variable is a random quantity following probability
distribution laws.
Regression Relationships
Terminology for two types of variables in econometrics:
Regression Relationships
or
E(Y |X1i , X2i , . . . , Xki ) = f (X1i , X2i , . . . , Xki ),
where
• Y : dependent variable
• Yi : specific actual value of the dependent variable
• X1 , X2 , . . . , Xk : independent variables
• X1i , X2i , . . . , Xki : specific values of the independent variables
• Ui : random error term corresponding to the i-th observation.
Regression Relationships
or
Ŷi = f (X1i , X2i , . . . , Xki ),
where ei is the error term in the sample, the residual, an estimate of
Ui .