Comparing Ordinary Least Squares, Ridge, and Lasso Regression for Multicollinearity Mitigation in Linear Models

  • F.A. Okolie Department of Mathematical Sciences, Ajayi Crowther University, Oyo State, Nigeria; Department of Statistics, Olabisi Onabanjo University, Ago-Iwoye, Ogun State, Nigeria
  • B.O. Fagbemigun Department of Mathematical Sciences, Ajayi Crowther University, Oyo State, Nigeria
  • O.E. Taiwo Department of Mathematical Sciences, Ajayi Crowther University, Oyo State, Nigeria
  • E.O. Adekola Department of Mathematical Sciences, Ajayi Crowther University, Oyo State, Nigeria
Keywords: Linear Modeling, Regression Analysis, Multicollinearity, Ridge Regression, LASSO Regression

Abstract

Ordinary Least Squares (OLS) regression provides unbiased estimates but performs poorly when predictor variables are highly correlated, due to increased variance and model instability. This study compares the effectiveness of OLS, Ridge regression, and the Least Absolute Shrinkage and Selection Operator (LASSO) in mitigating multicollinearity and improving predictive accuracy in linear models. Using academic data of University of Ilorin undergraduate students, Nigeria, we evaluated model performance using Root Mean Squared Error (RMSE), Variance Inflation Factors (VIF), and cross-validation. Ridge regression applies an L₂ penalty to shrink coefficients, while LASSO uses an L₁ penalty that also enables variable selection by setting some coefficients to zero. The results show that Ridge regression achieved the best generalization performance with the lowest test RMSE (0.2358), while LASSO provided a more interpretable model through coefficient sparsity. OLS exhibited overfitting and the poorest generalization due to high multicollinearity. The findings highlight the importance of regularization techniques in regression modeling, especially in high-dimensional data environments. This study offers practical guidance on model selection when predictive accuracy and feature interpretability are essential.an 

References

Montgomery, D. C., Peck, E. A., & Vining, G. G. (2021). Introduction to linear regression analysis. John Wiley & Sons.

Chan, J. Y., Leow, S. M., Bea, K. T., Cheng, W. K., Phoong, S. W., Hong, Z. W., & Chen, Y. L. (2022). Mitigating the multicollinearity problem and its machine learning approach: a review. Mathematics, 10(8), 1283.

Hoerl, A. E., & Kennard, R. W. (1970). Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12(1), 55-67.

Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1), 267-288.

Buhlmann, P., & Van De Geer, S. (2011). Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media.

Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2), 301-320.

Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American Statistical Association, 101(476), 1418-1429.

Fan, J., & Li, R. (2001). Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American Statistical Association, 96(456), 1348-1360.

Hastie, T., Tibshirani, R., & Wainwright, M. (2015). Statistical Learning with Sparsity: The Lasso and Generalizations. CRC Press.
Published
2025-12-10
How to Cite
Okolie, F., Fagbemigun, B., Taiwo, O., & Adekola, E. (2025). Comparing Ordinary Least Squares, Ridge, and Lasso Regression for Multicollinearity Mitigation in Linear Models. International Journal of Mathematical Sciences and Optimization: Theory and Applications, 11(4), 46 - 56. Retrieved from https://ijmso.unilag.edu.ng/article/view/2914
Section
Articles