Not Logged In

Reinforcement Learning for RoboCup-Soccer Keepaway

Full Text: AB05.pdf PDF

RoboCup simulated soccer presents many challenges to reinforcement learning methods, including a large state space, hidden and uncertain state, multiple independent agents learning simultaneously, and long and variable delays in the effects of actions. We describe our application of episodic SMDP Sarsa(λ) with linear tile-coding function approximation and variable λ to learning higher-level decisions in a keepaway subtask of RoboCup soccer. In keepaway, one team, “the keepers,” tries to keep control of the ball for as long as possible despite the efforts of “the takers.” The keepers learn individually when to hold the ball and when to pass to a teammate. Our agents learned policies that significantly outperform a range of benchmark policies. We demonstrate the generality of our approach by applying it to a number of task variations including different field sizes and different numbers of players on each team.

Citation

P. Stone, R. Sutton, G. Kuhlmann. "Reinforcement Learning for RoboCup-Soccer Keepaway". Adaptive Behavior, (13(3),), pp pp 165-188, March 2005.

Keywords: RoboCup, simultaneously, reinforcement, machine learning
Category: In Journal

BibTeX

@article{Stone+al:AdaptiveBehavior05,
  author = {Peter Stone and Richard S. Sutton and Gregory Kuhlmann},
  title = {Reinforcement Learning for RoboCup-Soccer Keepaway},
  Number = "13(3),",
  Pages = { pp 165-188},
  journal = {Adaptive Behavior},
  year = 2005,
}

Last Updated: May 31, 2007
Submitted by Staurt H. Johnson

University of Alberta Logo AICML Logo