Skip to content

Regularization

Regularization is a process of introducing additional information in order to solve an ill-posed problem or to prevent overfitting (Wikipedia Regularization).

Advantages for applying regularization to regression:

  • Preventing overfitting

  • Variable selection and removal of correlated variables

  • Converting ill-posed problems to well-posed by adding additional information via penalty parameter lambda

Variable selection methods:

  • Ridge method: shrink coefficients of correlated variables

  • LASSO (Least Absolute Shrinkage and Selection Operator) method: pick one variable and discard the others

  • Elastic Net Penalty: mixture of Ridge and LASSO methods