Aggregating Algorithm Regression

Aggregating Algorithm Regression (AAR) is an application of the Aggregating Algorithm to the online linear regression setting (see Vovk (2001)).

For the online linear regression problem with bounded signal {x:xtB}, Vovk (2001) proves that for the square-loss function AAR can achieve

{LTinfθ(LT(θ)+aθ2_)+nY2ln(TB2/a+1)},

where {LT(θ)} is the loss of any linear function of {x}.

The algorithm works as follows (all the vectors are columns):

Initialize vector b=0, and matrix A=aI;
FOR {t=1,2,s.}
Read new xtn
A=A+xtxtʹ
Output prediction γt=bʹA-1xt
Read new yt
b=b+ytxt
END FOR.

This algorithm almost repeats online Ridge Regression (RR), with the difference that in RR the matrix A is updated after the prediction has made. This means the method chooses to follow the best expert. Unfortunately, it there is no theoretical guarantees for RR in the same form as for AAR.

AAR adds an additional term to the RR minimization function: on the T+1 step its predictor θ achieves minθ(t=1T(θʹxt-yt)2+aθ2+(θʹxT+1)2), where a.

Bibliography

  • Vladimir Vovk. Competitive on-line statistics. International Statistical Review 69, 213–248 (2001).