Near-Optimal No-Regret Learning for General Convex Games

Published on ● Video Link: https://www.youtube.com/watch?v=9DA2w8CrBeA



Duration: 1:03:31
1,075 views
18


Gabriele Farina (Carnegie Mellon University)
https://simons.berkeley.edu/talks/near-optimal-no-regret-learning-general-convex-games
Structure of Constraints in Sequential Decision-Making

A recent line of work has established uncoupled learning dynamics such that, when employed by all players in a game, each player's regret after T repetitions grows polylogarithmically in T, an exponential improvement over the traditional guarantees within the no-regret framework. However, so far these results have only been limited to certain classes of games with structured strategy spaces---such as normal-form and extensive-form games. The question as to whether O(polylog T) regret bounds can be obtained for general convex and compact strategy sets---which occur in many fundamental models in economics and multiagent systems---while retaining efficient strategy updates is an important question. In this talk, we answer this in the positive by establishing the first uncoupled learning algorithm with O(log T) per-player regret in general convex games, that is, games with concave utility functions supported on arbitrary convex and compact strategy sets. Our learning dynamics are based on an instantiation of optimistic follow-the-regularized-leader over an appropriately lifted space using a self-concordant regularizer that is, peculiarly, not a barrier for the feasible region. Further, our learning dynamics are efficiently implementable given access to a proximal oracle for the convex strategy set, leading to O(loglog T) per-iteration complexity; we also give extensions when access to only a linear optimization oracle is assumed. Finally, we adapt our dynamics to guarantee O(sqrt(T)) regret in the adversarial regime. Even in those special cases where prior results apply, our algorithm improves over the state-of-the-art regret bounds either in terms of the dependence on the number of iterations or on the dimension of the strategy sets. Based on joint work with Ioannis Anagnostides, Haipeng Luo, Chung-Wei Lee, Christian Kroer, and Tuomas Sandholm. Paper link: https://arxiv.org/abs/2206.08742




Other Videos By Simons Institute for the Theory of Computing


2022-10-26Mathematics of the COVID-19 Pandemics: Lessons Learned and How to Mitigate the Next One
2022-10-25Efficient and Targeted COVID-19 Border Testing via Reinforcement Learning
2022-10-25Random Walks on Simplicial Complexes for Exploring Networks
2022-10-25Functional Law of Large Numbers and PDEs for Spatial Epidemic Models with...
2022-10-25Algorithms Using Local Graph Features to Predict Epidemics
2022-10-24Epidemic Models with Manual and Digital Contact Tracing
2022-10-21Pandora’s Box: Learning to Leverage Costly Information
2022-10-20Thresholds
2022-10-19NLTS Hamiltonians from Codes | Quantum Colloquium
2022-10-15Learning to Control Safety-Critical Systems
2022-10-14Near-Optimal No-Regret Learning for General Convex Games
2022-10-14The Power of Adaptivity in Representation Learning: From Meta-Learning to Federated Learning
2022-10-14When Matching Meets Batching: Optimal Multi-stage Algorithms and Applications
2022-10-13Optimal Learning for Structured Bandits
2022-10-13Dynamic Spatial Matching
2022-10-13New Results on Primal-Dual Algorithms for Online Allocation Problems With Applications to ...
2022-10-12Learning Across Bandits in High Dimension via Robust Statistics
2022-10-12Are Multicriteria MDPs Harder to Solve Than Single-Criteria MDPs?
2022-10-12A Game-Theoretic Approach to Offline Reinforcement Learning
2022-10-11The Statistical Complexity of Interactive Decision Making
2022-10-11A Tutorial on Finite-Sample Guarantees of Contractive Stochastic Approximation With...



Tags:
Simons Institute
theoretical computer science
UC Berkeley
Computer Science
Theory of Computation
Theory of Computing
Structure of Constraints in Sequential Decision-Making
Gabriele Farina