Project IV (MATH4072) 2017-18


Regularization

Dr J. Einbeck

Description

Most statistical estimation routines are, in some way, based on minimizing or maximizing a certain objective function, for instance least squares or Maximum Likelihood. These techniques hit a natural limit when the model possesses more parameters than data points, since invidual parameteters lose at this point their identifiability and standard estimation methods break down; for instance due to singularity of matrices that need to be inverted. While, from a traditional point of view, one may consider this problem (often referred to as the `p>n' problem) as a rather hypothetical consideration, modern data sets frequently show exactly this feature. For instance, genetic microarray data typically involve observations on some dozens of cells (observations) but thousands of genes (parameters). Hence, any statistical model dealing with such data is per se over-parametrized. Regularization is referred to as the process of dealing with overparametrized models by including a `penalty' term which enforces constraints on the parameters, making the estimation step feasible again. One of the most famous representatives of regularization methods is the `LASSO', which minimizes a LS criterrion with absolute-norm penalty,

(see also Reference 1. and 2. below), where λ is referred to as regularization (or penalty) parameter. The very useful feature of the LASSO is that it `shrinks' all parameters to 0 which are not really needed, hence carrying out implicit variable selection, enabling a sparse representation of the relationship between predictors and responses. Regularization appears in many other forms and facets, and can be considered as one of the current `hot' topics in Statistics and Machine Learning (Ref 3.). For instance, regularization is used in nonparametric regression where it enforces an adequate degree of smoothness of the regression function (see Ref. 4), or the analysis of multivariate time series (see Ref. 4).
In this project, you will focus on some special cases of regularization techniques which are accessible on 4H level, and apply them in the context of data sets of your interest.

Prerequisites

  • Statistical Methods III
  • Topics in Statistics III is useful but not necessary

Resources

  1. The Wikipedia page
  2. T. Hastie, R. Tibshirani, M. Wainwright, Statistical Learning with Sparsity
  3. Some interesting points of view on regularization in Machine Learning
  4. P. Eilers, B. Marx, and M. Durban, Twenty years of P-Splines
  5. C. Croux, Time series models: sparse estimation and robustness aspects

email: jochen.einbeck "at" durham.ac.uk