(Machine learning|Inverse problems) - Regularization

1 - About

Regularization refers to a process of introducing additional information in order to:

  • solve an ill-posed problem
  • or to prevent overfitting.

This information is usually of the form of a penalty for complexity, such as restrictions for smoothness or bounds on the vector space norm.

3 - Techniques

3.1 - Least Square

The least-squares method can be viewed as a very simple form of regularization.

3.2 - Linear Regression

In statistics and machine learning, regularization methods are used for model selection, in particular to prevent overfitting by penalizing models with extreme parameter values. The most common variants in machine learning are L₁ and L₂ regularization.

When applied in linear regression, the resulting models are termed ridge regression or lasso.

Statistics - (Shrinkage|Regularization) of Regression Coefficients

Regularization is also employed in:

3.3 - Bayes

From a Bayesian point of view, many regularization techniques correspond to imposing certain prior distributions on model parameters.

4 - Documentation / Reference

data_mining/regularization.txt · Last modified: 2014/09/14 22:00 by gerardnico