Online learning with random representations

作者: Richard S. Sutton , Steven D. Whitehead

DOI: 10.1016/B978-1-55860-307-3.50047-2

关键词:

摘要: We consider the requirements of online learning|learning which must be done incrementally and in realtime, with results learning available soon after each new example is acquired. Despite abundance methods for from examples, there are few that can used eectively learning, e.g., as components reinforcement systems. Most these few, including radial basis functions, CMACs, Kohonen’s self-organizing maps, those developed this paper, share same structure. All expand original input representation into a higher dimensional an unsupervised way, then map to nal answer using relatively simple supervised learner, such perceptron or LMS rule. Such structures learn very rapidly reliably, but have been thought either scale poorly require extensive domain knowledge. To contrary, some researchers (Rosenblatt, 1962; Gallant & Smith, 1987; Kanerva, 1988; Prager Fallside, 1988) argued expanded chosen largely at random good results. The main contribution paper develop test hypothesis. show random-representation perform well nearest-neighbor (while being more suited learning), signicantly better than backpropagation. nd size does increase dimensionality problem, not unreasonably so, required reduced substantially unsupervisedlearning techniques. Our suggest randomness has useful role play constructive induction. 1. Online Learning Applications divided two types: oine.

参考文章(32)
Jeffrey C Schlimmer, Richard H Granger, Incremental Learning from Noisy Data Machine Learning. ,vol. 1, pp. 317- 354 ,(1986) , 10.1023/A:1022810614389
TERENCE D. SANGER, OPTIMAL HIDDEN UNITS FOR TWO-LAYER NONLINEAR FEEDFORWARD NEURAL NETWORKS International Journal of Pattern Recognition and Artificial Intelligence. ,vol. 5, pp. 545- 561 ,(1991) , 10.1142/S0218001491000314
Eric Hartman, James D. Keeler, Predicting the future: Advantages of semilocal units Neural Computation. ,vol. 3, pp. 566- 578 ,(1991) , 10.1162/NECO.1991.3.4.566
Leonard Uhr, Charles Vossler, A pattern recognition program that generates, evaluates, and adjusts its own operators Papers presented at the May 9-11, 1961, western joint IRE-AIEE-ACM computer conference on - IRE-AIEE-ACM '61 (Western). pp. 555- 569 ,(1961) , 10.1145/1460690.1460751
R.W. Prager, F. Fallside, The modified Kanerva model for automatic speech recognition Computer Speech & Language. ,vol. 3, pp. 61- 81 ,(1989) , 10.1016/0885-2308(89)90015-6
Douglas H. Fisher, Knowledge acquisition via incremental conceptual clustering Machine Learning. ,vol. 2, pp. 139- 172 ,(1987) , 10.1023/A:1022852608280
Jerome H. Friedman, Multivariate Adaptive Regression Splines Annals of Statistics. ,vol. 19, pp. 1- 141 ,(1991) , 10.1214/AOS/1176347963
Gerald Tesauro, Practical Issues in Temporal Difference Learning Machine Learning. ,vol. 8, pp. 257- 277 ,(1992) , 10.1007/BF00992697
P. Földiák, Forming sparse representations by local anti-Hebbian learning Biological Cybernetics. ,vol. 64, pp. 165- 170 ,(1990) , 10.1007/BF02331346
Paul E. Utgoff, Incremental Induction of Decision Trees Machine Learning. ,vol. 4, pp. 161- 186 ,(1989) , 10.1023/A:1022699900025