Second-order non-stationary online learning for regression

作者: Edward Moroshko , Koby Crammer , Nina Vaits

DOI:

关键词:

摘要: The goal of a learner in standard online learning, is to have the cumulative loss not much larger compared with best-performing function from some fixed class. Numerous algorithms were shown this gap arbitrarily close zero, best that chosen off-line. Nevertheless, many real-world applications, such as adaptive filtering, are non-stationary nature, and prediction may drift over time. We introduce two novel for regression, designed work well environment. Our first algorithm performs resets forget history, while second last-step min-max optimal context drift. analyze both worst-case regret framework show they maintain an average slowly changing sequence linear functions, long sublinear. In addition, stationary case, when no occurs, our suffer logarithmic regret, previous algorithms. bounds improve existing ones, simulations demonstrate usefulness these other state-of-the-art approaches.

参考文章(38)
Nina Vaits, Koby Crammer, Re-adapting the Regularization of Weights for Non-stationary Regression Lecture Notes in Computer Science. pp. 114- 128 ,(2011) , 10.1007/978-3-642-24412-4_12
Manfred K. Warmuth, Peter Auer, Tracking the best disjunction Electronic Colloquium on Computational Complexity. ,vol. 7, ,(2000)
Eiji Takimoto, Manfred K. Warmuth, The Last-Step Minimax Algorithm Lecture Notes in Computer Science. pp. 279- 290 ,(2000) , 10.1007/3-540-40992-0_21
Katy S. Azoury, M. K. Warmuth, Relative loss bounds for on-line density estimation with the exponential family of distributions uncertainty in artificial intelligence. ,vol. 43, pp. 31- 40 ,(1999) , 10.1023/A:1010896012157
Matthew J. Streeter, H. Brendan McMahan, Adaptive Bound Optimization for Online Convex Optimization conference on learning theory. pp. 244- 256 ,(2010)
Jürgen Forster, On Relative Loss Bounds in Generalized Linear Regression fundamentals of computation theory. pp. 269- 280 ,(1999) , 10.1007/3-540-48321-7_22
B. WIDROW, M. E. HOFF, Adaptive switching circuits Neurocomputing: foundations of research. pp. 123- 134 ,(1988) , 10.21236/AD0241531
Nicolo Cesa-Bianchi, Manfred K. Warmuth, Philip M. Long, WORST-CASE QUADRATIC LOSS BOUNDS FOR ON-LINE PREDICTION OF LINEAR FUNCTIONS BY GRADIENT DESCENT University of California at Santa Cruz. ,(1993)
Steven Busuttil, Yuri Kalnishkan, Online Regression Competitive with Changing Predictors Lecture Notes in Computer Science. pp. 181- 195 ,(2007) , 10.1007/978-3-540-75225-7_17
Nicolo Cesa-Bianchi, Gabor Lugosi, Prediction, learning, and games ,(2006)