Not Logged In

Convergence and no-regret in multiagent learning

Full Text: 04nips.pdf PDF

Learning in a multiagent system is a challenging problem due to two key factors. First, if other agents are simultaneously learning then the environment is no longer stationary, thus undermining convergence guarantees. Second, learning is often susceptible to deception, where the other agents may be able to exploit a learner’s particular dynamics. In the worst case, this could result in poorer performance than if the agent was not learning at all. These challenges are identifiable in the two most common evaluation criteria for multiagent learning algorithms: convergence and regret. Algorithms focusing on convergence or regret in isolation are numerous. In this paper, we seek to address both criteria in a single algorithm by introducing GIGA-WoLF, a learning algorithm for normalform games. We prove the algorithm guarantees at most zero average regret, while demonstrating the algorithm converges in many situations of self-play. We prove convergence in a limited setting and give empirical results in a wider variety of situations. These results also suggest a third new learning criterion combining convergence and regret, which we call negative non-convergence regret (NNR).

Citation

M. Bowling. "Convergence and no-regret in multiagent learning". Neural Information Processing Systems (NIPS), Vancouver, British Columbia, Canada, pp 209-216, January 2005.

Keywords: machine learning
Category: In Conference

BibTeX

@incollection{Bowling:NIPS05,
  author = {Michael Bowling},
  title = {Convergence and no-regret in multiagent learning},
  Pages = {209-216},
  booktitle = {Neural Information Processing Systems (NIPS)},
  year = 2005,
}

Last Updated: April 24, 2007
Submitted by William Thorne

University of Alberta Logo AICML Logo