Uncertainty in Artificial Intelligence
First Name   Last Name   Password   Forgot Password   Log in!
    Proceedings   Proceeding details   Article details         Authors         Search    
Learning is planning: near Bayes-optimal reinforcement learning via Monte-Carlo tree search
John Asmuth, Michael Littman
Abstract:
Bayes-optimal behavior, while well-defined, is often difficult to achieve. Recent advances in the use of Monte-Carlo tree search (MCTS) have shown that it is possible to act near-optimally in Markov Decision Processes (MDPs) with very large or infinite state spaces. Bayes-optimal behavior in an unknown MDP is equivalent to optimal behavior in the known belief-space MDP, although the size of this belief-space MDP grows exponentially with the amount of history retained, and is potentially infinite. We show how an agent can use one particular MCTS algorithm, Forward Search Sparse Sampling (FSSS), in an efficient way to act nearly Bayes-optimally for all but a polynomial number of steps, assuming that FSSS can be used to act efficiently in any possible underlying MDP.
Keywords:
Pages: 19-26
PS Link:
PDF Link: /papers/11/p19-asmuth.pdf
BibTex:
@INPROCEEDINGS{Asmuth11,
AUTHOR = "John Asmuth and Michael Littman",
TITLE = "Learning is planning: near Bayes-optimal reinforcement learning via Monte-Carlo tree search",
BOOKTITLE = "Proceedings of the Twenty-Seventh Conference Annual Conference on Uncertainty in Artificial Intelligence (UAI-11)",
PUBLISHER = "AUAI Press",
ADDRESS = "Corvallis, Oregon",
YEAR = "2011",
PAGES = "19--26"
}


hosted by DSL   •   site info   •   help