In imperfect-information games, a common assumption is that players can perfectly model the strategic interaction and always maintain control over their decision points. We relax this assumption by introducing the notion of limited-control repeated games. In this setting, two players repeatedly play a zero-sum extensive-form game and, at each iteration, a player may lose control over portions of her game tree. Intuitively, this can be seen as the chance player hijacking the interaction and taking control of certain decision points. What subsequently happens is no longer controllable-or even known-by the original players. We introduce pruned fictitious play, a variation of fictitious play that can be employed by the players to reach an equilibrium in limited-control repeated games. We motivate this technique with the notion of limited best response, which is the key step of the learning rule we employ. We provide a general result on the probabilistic guarantees of a limited best response with respect to the original game model. Then, we experimentally evaluate our technique and show that pruned fictitious play has good convergence properties.

Learning dynamics in limited-control repeated games

Celli, Andrea;
2018

Abstract

In imperfect-information games, a common assumption is that players can perfectly model the strategic interaction and always maintain control over their decision points. We relax this assumption by introducing the notion of limited-control repeated games. In this setting, two players repeatedly play a zero-sum extensive-form game and, at each iteration, a player may lose control over portions of her game tree. Intuitively, this can be seen as the chance player hijacking the interaction and taking control of certain decision points. What subsequently happens is no longer controllable-or even known-by the original players. We introduce pruned fictitious play, a variation of fictitious play that can be employed by the players to reach an equilibrium in limited-control repeated games. We motivate this technique with the notion of limited best response, which is the key step of the learning rule we employ. We provide a general result on the probabilistic guarantees of a limited best response with respect to the original game model. Then, we experimentally evaluate our technique and show that pruned fictitious play has good convergence properties.
2018
2019
Celli, Andrea; Marchesi, Alberto
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11565/4043611
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact