Proposal-Ridge Regression
Proposal-Ridge Regression
.
Summary
The choice of ridge parameter plays a pivotal role in achieving the desirable property i.e. the
smaller variance. Generally, this is achieved at the cost of introducing a bias in the estimates.
In this work, we will propose a new method for selection of optimal value of ridge parameter.
The main objective in this new proposal is to achieve the smaller variance without a
substantial increase in the bias. The proposed estimators will be compared with the ordinary
least square and existing ridge estimators. Extensive simulations will be performed to
evaluate the performance for small to large samples. Real-life examples will be provided to
illustrate the use of proposed methods.
1. Introduction
and
( ̂)
The OLS estimator is unbiased and may have smaller variance under usual conditions
but its covariance matrix is heavily dependent upon the characteristics of matrix.
When the independent variables are correlated, the OLS remains unbiased but it leads to a
substantial increase in the variance.
1
To overcome the problem of multicollinearity, Hoerl and Kennard (1970) proposed ridge
regression (RR) estimator by introducing a biasing parameter, , in the objective function of
the Lagrange multiplier as follows:
where represents a constant value, i.e., . The normal equations for regression
parameter are obtained as:
( )
̂ ( )
where represent the biasing parameter, denotes the number of independent variables and
denotes the identity matrix of order . The biasing parameter plays a vital role in
tradeoff between the bias and consistency of the estimator. When ̂ approaches to
OLS estimator and we obtain the unbiased estimator but inconsistent estimator of . Further,
when , ̂ we obtain consistent, stable but biased estimator of For positive
value of biasing parameter ( ), the estimator provides smaller mean squared error compared
to the OLS estimator (Muniz & Kibria, 2009).
2
values obtained from the matrix. The OLS estimator ( ̂), and the ridge regression
estimator ̂ in the canonical form are defined as follows:
̂
̂ ∑( ⁄ )
̂ ∑ ∑
To deal with the problem of multicollinearity, Hoerl and Kennard (1970) introduced an
alternative technique to OLS using (4). To use ridge regression, a researcher needs to
determine the value of biasing parameter ( ). Thus, for the estimation of many methods
have been recommended by the investigators in terms of minimum MSE of ridge estimator
which depends on parameter vector of regression coefficient ( ) and the variance of MLR
model ( ). Since and are unknown, the researcher needs to employ their estimates. In
past twenty years, the researchers have been focusing on the estimation of biasing parameter
( ) by several ways and under different situations and then compare the results with the OLS
estimator. Much discussion concerning ridge regression go around finding an optimal value
of biasing parameter .
To estimate the biasing parameter ( ), Hoerl and Kennard (1970) defined the ridge
estimator as follows:
̂ ̂ ( )
̂
where ̂ represents the residual mean square estimate and can be computed using ̂
̂
∑ , ̂ are the OLS residuals and ̂ (̂ ̂ ).
3
̂
̂
∑ ̂
Some new methods to estimate were suggested by Kibria (2003) by taking the
arithmetic mean, geometric mean and median of ̂ ⁄ ̂ respectively are defined as:
̂
̂ ∑( )
̂
̂
̂
√∏ ̂
̂
̂ ( )
̂
In general, biasing parameter ( ) in ridge regression leads to reduced variance but the cost
may be high in terms of bias. To deal with this problem, we will propose some new
estimators that may achieve smaller MSE with no or very little bias.
4. Research Objectives:
i. To propose a new ridge estimator to address the problem of multicollinearity with smaller
variance and bias.
5. Methodology:
To evaluate the performance of the proposed estimators in this study, the MSE criterions will
be used. Many researchers e.g., Kibria (2003), Suhail et al., (2020) have used the MSE
criterion for performance analysis of estimators. The MSE is described as follows:
̂ ̂ ̂ ∑ ̂
6. Proposed Analysis:
4
Following Kibria (2003) and Suhail et al., (2021), we have generated the independent
variables equation as follows:
where represents the specified degree of correlation between the independent variables
and is generated using the standard normal distribution. The response variable is obtained
by:
̂ ∑ ̂ ̂
References
Dorugade, A., 2014. New ridge parameters for ridge regression. Journal of the Association of
Arab Universities for Basic and Applied Sciences, 15(1), pp. 94-99.
Hoerl, A. E. & Kennard, R. W., 1970. Ridge regression: Biased estimation for nonorthogonal
problems. Technometrics, 12(1), pp. 55-67.
5
Hoerl, A. E., Kennard, R. W. & Baldwin, K. F., 1975. Ridge regression: Some simulations.
Communications in Statistics-Theory and Methods, 4(2), pp. 105-123.
Khalaf, G., Mansson, K. & Shukur, G., 2013. Modified ridge regression estimators.
Communications in Statistics-Theory and Methods, 42(8), pp. 1476-1487.
Kibria, B. G., 2003. Performance of some new ridge regression estimators. Communications
in Statistics-Simulation and Computation, 32(2), pp. 419-435.
Kibria, B. & Lukman, A., 2020. A new ridge-type estimator for the linear regression model:
Simulations and applications. Scientifica, Volume 2020, pp. 1-16.
Muniz, G. & Kibria, B. G., 2009. On some ridge regression estimators: An empirical
comparison. Communications in Statistics - Simulation and Computation, 38(3), pp. 621-630.
Suhail, M., Chand, S. & Aslam, M., 2021. New quantile based ridge M-estimator for linear
regression models with multicollinearity and outliers. Communications in Statistics-
Simulation and Computation, pp. 1-25.
Suhail, M., Chand, S. & Kibria, . B. M. G., 2020. Quantile based estimation of biasing
parameters in ridge regression model. Communications in Statistics-Simulation and
Computation, 49(10), pp. 2732-2744.
Suhail, M., Chand, S. & Kibria, G. B., 2019. Quantile-based robust ridge m-estimator for
linear regression model in presence of multicollinearity and outliers. Communications in
Statistics-Simulation and Computation, pp. 1-13.