Chapter 14 - Nonlinear Regression Models
Chapter 14 - Nonlinear Regression Models
This chapter covers models that are nonlinear in parameters (whether the variables are linear or
not)
Intrinsically linear and intrinsically
nonlinear models
Some models appear nonlinear in the parameters but are inherently / intrinsically linear
because with a suitable transformation they can be converted to linear-in-the-parameters
models.
If nonlinear models cannot be linearized in their parameters, they are intrinsically nonlinear
regression models.
When referring to NLRM in this chapter = Intrinsically nonlinear
Examples 1 – Nonlinear in parameters?
Really?
What if we let ?
ln 𝑌 𝑖 =𝛼+ β 2 ln X i +ui
Model is linear in and
Examples 2 - Nonlinear in parameters?
Really??
ln both sides:
ln 𝑌 𝑖 =𝛽 1+ 𝛽 2 𝑋 𝑖 +𝑢𝑖
Now?
ln
( 1− 𝑌 𝑖
𝑌𝑖 )= 𝛽1 + 𝛽2 𝑋 𝑖 +𝑢𝑖
Estimation of linear and nonlinear
regressions
𝒀 𝒊= 𝜷𝟏 + 𝜷𝟐 𝑿 𝒊 +𝒖𝒊 𝒀 𝒊= 𝜷𝟏 𝒆 𝜷 𝟐 𝑿
+𝒖 𝒊
Estimate by using OLS (minimize RSS) Estimate by using OLS (minimize RSS)
Explicit solutions are obtained of the 2 Cannot obtain explicit solutions of unknowns
unknowns in terms of the data (unknowns on (unknowns both on left-hand and right-hand
the left-hand side and knowns on the right- side of equation)
hand side)
Unknowns are expressed in terms of
themselves and the data
Estimation of linear and nonlinear
regressions
We can therefore apply OLS to estimate NLRM parameters but we can’t obtain explicit solutions
of the unknowns.
OLS applied to NLRM is known as nonlinear least squares
What is the solution?
Estimating NLRM: trial and error
method
Example
◦ Relationship between management fees paid to investment advisors to manage assets and the net asset
value of the fund
Example (continued)
Example continued
Make use of trial and error / iterative process to estimate the model (exponential model)
Initially use and (pure guesses / prior experience)
We have enough info to calculate error sum of squares
With OLS objective is to find values of unknown parameters that will minimize error sum of
squares (this happens if estimated Y values are as close as possible to actual Y values).
How do we know that this is the minimum?
Repeat the process but use and
…..much smaller!
Example continued
Trial and error is an iterative and time-consuming process!
We need an algorithm to help with the process
Approaches to estimating NLRM
Direct search / trial and error
Direct optimization
Iterative linearization
Direct search / trial and error /
derivative-free method
As example covered – this method is generally not used
Disadvantages
◦ If NLRM involves several parameters, method becomes cumbersome and computationally expensive
◦ No guarantee that final set of parameter values selected, will necessarily give absolute minimum error
sum of squares
Direct optimization
Error sum of squares are differentiated with respect to each unknown coefficient, set resulting
equation to zero and solve resulting normal equations simultaneously
These equations cannot be solved explicitly / analytically
Therefore need an iterative (again!) routine
Possible routine: Method of steepest descent
◦ Very involved method
◦ Also requires initial values but proceeds more systematically than trial and error method
◦ Disadvantage: May converge to final values of parameters extremely slowly
Iterative linearization method
Nonlinear equation is linearized around some initial values of the parameters
Linearized equation is then estimated by using OLS and initially chosen values are adjusted
Adjusted values used to relinearize the model and estimated again by using OLS and readjust
estimated values
Process is continued until there is no substantial change in estimated values from last couple of
iterations
Main technique used in linearizing nonlinear equation – Taylor series expansion (calculus)
Estimating NLRM using Taylor series expansion systematized into 2 algorithms
◦ Gauss-Newton iterative method
◦ Newton-Raphson iterative method
Eviews
Quadratic hill climbing
Newton-Raphson
Berndt-Hall-Hall-Hausman
The least-squares estimators for nonlinear regression models with normal error terms, when the
sample size is large, are approximately normally distributed and almost unbiased and have
almost minimum variance. The large-sample theory also applies when the error terms are not
normally distributed.
Therefore t-statistics and the F-stat are meaningful only if interpreted in the large-sample
context.
Practical this week
No practical application on Chapter 14
Next week…..
Chapter 15 – Part 1.
Remember that you are writing a class test on Thursday 25 July 2024 at 13:00 in D1 lab 108
about the practical application of Chapters 1 to 13 of Gujarati (everything we covered in
Econometrics 3A).