Equilibrium Finding in Normal-Form Games Via Greedy Regret Minimization

Hugh Zhang, Adam Lerer, Noam Brown

We extend the classic regret minimization framework for approximating equilibria in normal-form games by greedily weighing iterates based on regrets observed at runtime. Theoretically, our method retains all previous convergence rate guarantees. Empirically, experiments on large randomly generated games and normal-form subgames of the AI benchmark Diplomacy show that greedy weights outperforms previous methods whenever sampling is used, sometimes by several orders of magnitude.

Knowledge Graph

arrow_drop_up

Comments

Sign up or login to leave a comment