Not Logged In

Rational and Convergent Learning in Stochastic Games

Full Text: bowling01rational.pdf PDF

This paper investigates the problem of policy learning in multiagent environments using the stochastic game framework, which we briefly overview. We introduce two properties as desirable for a learning agent when in the presence of other learning agents, namely rationality and convergence. We examine existing reinforcement learning algorithms according to these two properties and notice that they fail to simultaneously meet both criteria. We then contribute a new learning algorithm,WoLF policy hillclimbing, that is based on a simple principle: “learn quickly while losing, slowly while winning.” The algorithm is proven to be rational and we present empirical results for a number of stochastic games showing the algorithm converges.

Citation

M. Bowling, M. Veloso. "Rational and Convergent Learning in Stochastic Games". International Joint Conference on Artificial Intelligence (IJCAI), pp 1021-1026, August 2001.

Keywords:  
Category: In Conference

BibTeX

@incollection{Bowling+Veloso:IJCAI01,
  author = {Michael Bowling and Manuela Veloso},
  title = {Rational and Convergent Learning in Stochastic Games},
  Pages = {1021-1026},
  booktitle = {International Joint Conference on Artificial Intelligence
    (IJCAI)},
  year = 2001,
}

Last Updated: April 24, 2007
Submitted by AICML Admin Assistant

University of Alberta Logo AICML Logo