Mathematical Formulation: Ridge Regression Estimator
Consider the multiple linear regression model:
[Tex]y=Xβ+ϵ[/Tex]
where y is an n×1 vector of observations, X is an [Tex]n*p[/Tex] matrix of predictors, β is a p×1 vector of unknown regression coefficients, and ϵ is an n×1 vector of random errors. The OLS estimator of β is given by:
[Tex]\hat{\beta}_{\text{OLS}} = (X’X)^{-1}X’y [/Tex]
In the presence of multicollinearity, [Tex]X^′X[/Tex] is nearly singular, leading to unstable estimates. The ridge regression estimator modifies this by introducing the ridge parameter k:
[Tex]\hat{\beta}_k = (X’X + kI)^{-1}X’y [/Tex]
What is Ridge Regression?
Ridge regression, also known as Tikhonov regularization, is a technique used in linear regression to address the problem of multicollinearity among predictor variables. Multicollinearity occurs when independent variables in a regression model are highly correlated, which can lead to unreliable and unstable estimates of regression coefficients.
Ridge regression mitigates this issue by adding a regularization term to the ordinary least squares (OLS) objective function, which penalizes large coefficients and thus reduces their variance.
Table of Content
- What is Ridge Regression?
- Mathematical Formulation: Ridge Regression Estimator
- Bias-Variance Tradeoff for Ridge Regression
- Selection of the Ridge Parameter in Ridge Regression
- 1. Cross-Validation
- 2. Generalized Cross-Validation (GCV)
- 3. Information Criteria
- 4. Empirical Bayes Methods
- 5. Stability Selection
- Practical Considerations for Selecting Ridge Parameter
- Use Cases and Applications of Ridge Regression
- Advantages and Disadvantages of Ridge Regression