# Solved – Understanding negative ridge regression

I'm looking for literature about negative ridge regression.

In short, it is a generalization of linear ridge regression using negative \$lambda\$ in the estimator formula: \$\$hatbeta = ( X^top X + lambda I)^{-1} X^top y.\$\$ The positive case has a nice theory: as a loss function, as a constraint, as a Bayes prior… but I feel lost with the negative version with only the above formula. It happens to be useful for what I am doing but I fail to interpret it clearly.

Do you know any serious introductory text about negative ridge? How can it be interpreted?

Contents

Here is a geometric illustration of what is going on with negative ridge.

I will consider estimators of the form \$\$hat{boldsymbolbeta}_lambda = (mathbf X^top mathbf X + lambda mathbf I)^{-1}mathbf X^topmathbf y\$\$ arising from the loss function \$\$mathcal L_lambda = |mathbf y – mathbf Xboldsymbolbeta|^2 + lambda |boldsymbolbeta|^2.\$\$ Here is a rather standard illustration of what happens in a two-dimensional case with \$lambdain[0,infty)\$. Zero lambda corresponds to the OLS solution, infinite lambda shrinks the estimated beta to zero:

Now consider what happens when \$lambdain(-infty, -s^2_max)\$, where \$s_mathrm{max}\$ is the largest singular value of \$mathbf X\$. For very large negative lambdas, \$hat{boldsymbolbeta}_lambda\$ is of course close to zero. When lambda approaches \$-s^2_max\$, the term \$(mathbf X^top mathbf X + lambda mathbf I)\$ gets one singular value approaching zero, meaning that the inverse has one singular value going to minus infinity. This singular value corresponds to the first principal component of \$mathbf X\$, so in the limit one gets \$hat{boldsymbolbeta}_lambda\$ pointing in the direction of PC1 but with absolute value growing to infinity.

What is really nice, is that one can draw it on the same figure in the same way: betas are given by points where circles touch the ellipses from the inside:

When \$lambdain(-s^2_mathrm{min},0]\$, a similar logic applies, allowing to continue the ridge path on the other side of the OLS estimator. Now the circles touch the ellipses from the outside. In the limit, betas approach the PC2 direction (but it happens far outside this sketch):

The \$(-s^2_mathrm{max}, -s^2_mathrm{min})\$ range is something of an energy gap: estimators there do not live on the same curve.

UPDATE: In the comments @MartinL explains that for \$lambda<-s^2_mathrm{max}\$ the loss \$mathcal L_lambda\$ does not have a minimum but has a maximum. And this maximum is given by \$hat{boldsymbolbeta}_lambda\$. This is why the same geometric construction with the circle/ellipse touching keeps working: we are still looking for zero-gradient points. When \$-s^2_mathrm{min}<lambdale 0\$, the loss \$mathcal L_lambda\$ does have a minimum and it is given by \$hat{boldsymbolbeta}_lambda\$, exactly as in the normal \$lambda>0\$ case.

But when \$-s^2_mathrm{max}<lambda<-s^2_mathrm{min}\$, the loss \$mathcal L_lambda\$ does not have either maximum or minimum; \$hat{boldsymbolbeta}_lambda\$ would correspond to a saddle point. This explains the "energy gap".

The \$lambdain(-infty, -s^2_max)\$ naturally arises from a particular constrained ridge regression, see The limit of "unit-variance" ridge regression estimator when \$lambdatoinfty\$. This is related to what is known in the chemometrics literature as "continuum regression", see my answer in the linked thread.

The \$lambdain(-s^2_mathrm{min},0]\$ can be treated in exactly the same way as \$lambda>0\$: the loss function stays the same and the ridge estimator provides its minimum.

Rate this post