Publication Type
Conference Proceeding Article
Version
publishedVersion
Publication Date
2-2024
Abstract
Many real-world scenarios including fleet management and Ad auctions can be modeled as Stackelberg mean-field games (SMFGs) where a leader aims to incentivize a large number of homogeneous self-interested followers to maximize her utility. Existing works focus on cases with a small number of heterogeneous followers, e.g., 5-10, and suffer from scalability issue when the number of followers increases. There are three major challenges in solving large-scale SMFGs: i) classical methods based on solving differential equations fail as they require exact dynamics parameters, ii) learning by interacting with environment is data-inefficient, and iii) complex interaction between the leader and followers makes the learning performance unstable. We address these challenges through transition-informed reinforcement learning. Our main contributions are threefold: i) we first propose an RL framework, the Stackelberg mean-field update, to learn the leader's policy without priors of the environment, ii) to improve the data efficiency and accelerate the learning process, we then propose the Transition-Informed Reinforcement Learning (TIRL) by leveraging the instantiated empirical Fokker-Planck equation, and iii) we develop a regularized TIRL by employing various regularizers to alleviate the sensitivity of the learning performance to the initialization of the leader's policy. Extensive experiments on fleet management and food gathering demonstrate that our approach can scale up to 100,000 followers and significantly outperform existing baselines.
Keywords
Multiagent Learning, Reinforcement Learning
Discipline
Artificial Intelligence and Robotics | Numerical Analysis and Scientific Computing
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
Proceedings of the 38th AAAI Conference on Artificial Intelligence (AAAI) 2024: Vancouver, May 6-10
First Page
17469
Last Page
17476
Identifier
10.1609/aaai.v38i16.29696
Publisher
AAAI Press
City or Country
Washington, DC
Citation
LI, Pengdeng; YU, Runsheng; WANG, Xinrun; and AN, Bo.
Transition-informed reinforcement learning for large-scale Stackelberg mean-field games.. (2024). Proceedings of the 38th AAAI Conference on Artificial Intelligence (AAAI) 2024: Vancouver, May 6-10. 17469-17476.
Available at: https://ink.library.smu.edu.sg/sis_research/9127
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1609/aaai.v38i16.29696
Included in
Artificial Intelligence and Robotics Commons, Numerical Analysis and Scientific Computing Commons