La régularisation de Tikhonov et la régression de crête sont des termes souvent utilisés comme s'ils étaient identiques. Est-il possible de spécifier exactement quelle est la différence?
La régularisation de Tikhonov et la régression de crête sont des termes souvent utilisés comme s'ils étaient identiques. Est-il possible de spécifier exactement quelle est la différence?
La régularisation de Tikhonov est un ensemble plus vaste que la régression en arête. Voici ma tentative pour expliquer exactement comment ils diffèrent.
Supposons que pour une matrice connue et un vecteur b , nous souhaitons trouver un vecteur x tel que:
.
L'approche standard est la régression linéaire par les moindres carrés ordinaires. Cependant, si aucun ne satisfait l'équation ou si plus d'un x le fait, la solution n'est pas unique, le problème est dit mal posé. Les moindres carrés ordinaires cherchent à minimiser la somme des résidus au carré, ce qui peut être écrit de manière compacte comme suit:
où est la norme euclidienne. En notation matricielle la solution, notée x , est donnée par:
Tikhonov regularization minimizes
for some suitably chosen Tikhonov matrix, . An explicit matrix form solution, denoted by , is given by:
this reduces to the unregularized least squares solution provided that (ATA)−1 exists.
Typically for ridge regression, two departures from Tikhonov regularization are described. First, the Tikhonov matrix is replaced by a multiple of the identity matrix
,
leading to
Finally, for ridge regression, it is typically assumed that variables are scaled so that has the form of a correlation matrix. and is the correlation vector between the variables and , leading to
Note in this form the Lagrange multiplier is usually replaced by , , or some other symbol but retains the property
In formulating this answer, I acknowledge borrowing liberally from Wikipedia and from Ridge estimation of transfer function weights
Carl has given a thorough answer that nicely explains the mathematical differences between Tikhonov regularization vs. ridge regression. Inspired by the historical discussion here, I thought it might be useful to add a short example demonstrating how the more general Tikhonov framework can be useful.
First a brief note on context. Ridge regression arose in statistics, and while regularization is now widespread in statistics & machine learning, Tikhonov's approach was originally motivated by inverse problems arising in model-based data assimilation (particularly in geophysics). The simplified example below is in this category (more complex versions are used for paleoclimate reconstructions).
Imagine we want to reconstruct temperaturesu[x,t=0] in the past, based on present-day measurements u[x,t=T] . In our simplified model we will assume that temperature evolves according to the heat equation
Tikhonov regularization can solve this problem by solving
Below is a comparison of the results:
We can see that the original temperatureu0 has a smooth profile, which is smoothed still further by diffusion to give ufwd . Direct inversion fails to recover u0 , and the solution uinv shows strong "checkerboarding" artifacts. However the Tikhonov solution ureg is able to recover u0 with quite good accuracy.
Note that in this example, ridge regression would always push our solution towards an "ice age" (i.e. uniform zero temperatures). Tikhonov regression allows us a more flexible physically-based prior constraint: Here our penalty essentially says the reconstructionu should be only slowly evolving, i.e. ut≈0 .
Matlab code for the example is below (can be run online here).
la source