Dinh, Le Cong, Nguyen, Tri-Dung, Zemhoho, Alain B., Tran-Thanh, Long (2021) Last round convergence and no-dynamic regret in asymmetric repeated games. In: Proceedings of Machine Learning Research. Proceedings of the 32nd International Conference on Algorithmic Learning Theory. 132. pp. 553-577. PMLR (The full text of this publication is not currently available from this repository. You may be able to access a copy if URLs are provided) (KAR id:93636)
The full text of this publication is not currently available from this repository. You may be able to access a copy if URLs are provided. (Contact us about this Publication) | |
Official URL: http://algorithmiclearningtheory.org/alt2021/?mscl... |
Abstract
This paper considers repeated games in which one player has a different objective than others. In particular, we investigate repeated two-player zero-sum games where the column player not only aims to minimize her regret but also stabilize the actions. Suppose that while repeatedly playing this game, the row player chooses her strategy at each round by using a no-regret algorithm to minimize her regret. We develop a no-dynamic regret algorithm for the column player to exhibit last round convergence to a minimax equilibrium. We show that our algorithm is efficient against a large set of popular no-regret algorithms the row player can use, including the multiplicative weights update algorithm, general follow-the-regularized-leader and any no-regret algorithms satisfy a property so called ?stability?.
Item Type: | Conference or workshop item (Paper) |
---|---|
Subjects: | H Social Sciences |
Divisions: | Divisions > Kent Business School - Division > Department of Analytics, Operations and Systems |
Depositing User: | Tri-Dung Nguyen |
Date Deposited: | 17 Mar 2022 14:58 UTC |
Last Modified: | 05 Nov 2024 12:58 UTC |
Resource URI: | https://kar.kent.ac.uk/id/eprint/93636 (The current URI for this page, for reference purposes) |
- Export to:
- RefWorks
- EPrints3 XML
- BibTeX
- CSV
- Depositors only (login required):