Paper on RL, exploration, hidden state

Andrew McCallum mccallum at cs.rochester.edu
Mon Jul 1 19:29:55 EDT 1996


The following paper on reinforcement learning, hidden state and
exploration is available by FTP.  Comments and suggestions are
welcome.

 "Efficient Exploration in Reinforcement Learning with Hidden State"

		       Andrew Kachites McCallum

			(submitted to NIPS)

			       Abstract

  Undoubtedly, efficient exploration is crucial for the success of a
  learning agent.  Previous approaches to directed exploration in
  reinforcement learning exclusively address exploration in Markovian
  domains, i.e. domains in which the state of the environment is fully
  observable.  If the environment is only partially observable, they
  cease to work because exploration statistics are confounded between
  aliased world states.

  This paper presents Fringe Exploration, a technique for efficient
  exploration in partially observable domains.  The key idea,
  (applicable to many exploration techniques), is to keep statistics
  in the space of possible short-term memories, instead of in the
  agent's current state space.  Experimental results in a partially
  observable maze and in a difficult driving task with visual routines
  show dramatic performance improvements.
 

Retrieval information:

FTP-host:      ftp.cs.rochester.edu
FTP-pathname:  /pub/papers/robotics/96.mccallum-nips.ps.gz
URL: ftp://ftp.cs.rochester.edu/pub/papers/robotics/96.mccallum-nips.ps.gz


More information about the Connectionists mailing list