New RL paper and WWW interface to archive

Rich Sutton sutton at gte.com
Fri May 26 18:26:12 EDT 1995


                GENERALIZATION IN REINFORCEMENT LEARNING:
             SUCCESSFUL EXAMPLES USING SPARSE COARSE CODING
                          Richard S. Sutton
                         submitted to NIPS'95

On large problems, reinforcement learning systems must use
parameterized function approximators such as neural networks in order to
generalize between similar situations and actions. In these cases there
are no strong theoretical results on the accuracy of convergence, and
computational results have been mixed. In particular, Boyan and Moore
reported at last year's meeting a series of negative results in
attempting to apply dynamic programming together with function approximation
to simple control problems with continuous state spaces. In this paper,
we present positive results for all the control tasks they attempted,
and for one that is significantly larger. The most important differences
are that we used sparse-coarse-coded function approximators (CMACs)
whereas they used mostly global function approximators, and that we
learned online whereas they learned offline. Boyan and Moore and
others have suggested that the problems they encountered could be solved
by using actual outcomes ("rollouts"), as in classical Monte Carlo
methods, and as in the TD(lambda) algorithm when lambda=1.
However, in our experiments this always resulted in substantially poorer
performance. We conclude that reinforcement learning can work robustly
in conjunction with function approximators, and that there is little
justification at present for avoiding the case of general lambda.

________________
The paper is available by ftp as
    ftp://ftp.cs.umass.edu/pub/anw/pub/sutton/sutton-inprer.ps.gz
or via a new WWW interface to my small archive at
    http://envy.cs.umass.edu/People/sutton/archive.html.
Please change any WWW pointers to the old ftp archive.




More information about the Connectionists mailing list