Abstract
The Markov game is a popular reinforcement learning framework for modeling competitive players in a dynamic environment. However, most of the existing works on Markov games focus on computing a certain equilibrium following uncertain interactions among the players but ignore the uncertainty of the environment model, which is ubiquitous in practical scenarios. In this work, we develop a theoretical solution to Markov games with environment model uncertainty. Specifically, we propose a new and tractable notion of robust correlated equilibria for Markov games with environment model uncertainty. In particular, we prove that the robust correlated equilibrium has a simple modification structure, and its characterization of equilibria critically depends on the environment model uncertainty. Moreover, we propose the first fully-decentralized stochastic algorithm for computing such the robust correlated equilibrium. Our analysis proves that the algorithm achieves the polynomial episode complexity Oe(SA2H5∊−2) for computing an approximate robust correlated equilibrium with ∊ accuracy.
| Original language | English |
|---|---|
| Journal | Journal of Machine Learning Research |
| Volume | 24 |
| State | Published - 2023 |
Keywords
- model uncertainty
- reinforcement learning
- robust Markov games
- robust correlated equilibrium
Fingerprint
Dive into the research topics of 'Decentralized Robust V-learning for Solving Markov Games with Model Uncertainty'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver