Not Logged In

Safe strategies for agent modelling in games

Full Text: fdsfs.pdf PDF

Research in opponent modelling has shown success, but a fundamental question has been overlooked: what happens when a modeller is faced with an opponent that cannot be successfully modelled? Many opponent modellers could do arbitrarily poorly against such an opponent. In this paper, we aim to augment opponent modelling techniques with a method that enables models to be used safely. We introduce -safe strategies, which bound by  the possible loss versus a safe value. We also introduce the Safe Policy Selection algorithm (SPS) as a method to vary  in a controlled fashion. We prove in the limit that an agent using SPS is guaranteed to attain at least a safety value in the cases when the opponent modelling is ineffective. We also show empirical evidence that SPS does not adversely affect agents that are capable of modelling the opponent. Tests with a domain of complicated modellers show that SPS is effective at eliminating losses while retaining wins in a variety of modelling algorithms.

Citation

P. McCracken, M. Bowling. "Safe strategies for agent modelling in games". National Conference on Artificial Intelligence (AAAI), San Jose, California, USA, October 2004.

Keywords: machine learning
Category: In Conference

BibTeX

@incollection{McCracken+Bowling:AAAI04,
  author = {Peter McCracken and Michael Bowling},
  title = {Safe strategies for agent modelling in games},
  booktitle = {National Conference on Artificial Intelligence (AAAI)},
  year = 2004,
}

Last Updated: April 24, 2007
Submitted by AICML Admin Assistant

University of Alberta Logo AICML Logo