no code implementations • 12 Feb 2024 • Yuepeng Yang, Antares Chen, Lorenzo Orecchia, Cong Ma
On the analytical front, we provide a refined $\ell_\infty$ error analysis of the weighted MLE that is more explicit and tighter than existing analyses.
no code implementations • 26 Sep 2022 • Yuepeng Yang, Cong Ma
We prove that optimistic-follow-the-regularized-leader (OFTRL), together with smooth value updates, finds an $O(T^{-1})$-approximate Nash equilibrium in $T$ iterations for two-player zero-sum Markov games with full information.