Publication Type
Conference Proceeding Article
Version
acceptedVersion
Publication Date
11-2024
Abstract
Public health practitioners often have the goal of monitoring patients and maximizing patients’ time spent in “favorable” or healthy states while being constrained to using limited resources. Restless multi-armed bandits (RMAB) are an effective model to solve this problem as they are helpful to allocate limited resources among many agents under resource constraints, where patients behave differently depending on whether they are intervened on or not. However, RMABs assume the reward function is known. This is unrealistic in many public health settings because patients face unique challenges and it is impossible for a human to know who is most deserving of any intervention at such a large scale. To address this shortcoming, this paper is the first to present the use of inverse reinforcement learning (IRL) to learn desired rewards for RMABs, and we demonstrate improved outcomes in a maternal and child health telehealth program. First we allow public health experts to specify their goals at an aggregate or population level and propose an algorithm to design expert trajectories at scale based on those goals. Second, our algorithm WHIRL uses gradient updates to optimize the objective, allowing for efficient and accurate learning of RMAB rewards. Third, we compare with existing baselines and outperform those in terms of run-time and accuracy. Finally, we evaluate and show the usefulness of WHIRL on thousands on beneficiaries from a real-world maternal and child health setting in India.
Discipline
Health Information Technology | Operations Research, Systems Engineering and Industrial Engineering | Theory and Algorithms
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
PRICAI 2024: Trends in artificial intelligence: 21st Pacific Rim International Conference on Artificial Intelligence, Kyoto, November 18-24
Volume
15285
First Page
165
Last Page
178
ISBN
9789819601271
Identifier
10.1007/978-981-96-0128-8_15
Publisher
Springer
City or Country
Cham
Citation
JAIN, Gauri; VARAKANTHAM, Pradeep; XU, Haifeng; TANEJA, Aparna; DOSHI, Prashant; and TAMBE, Milind.
IRL for restless multi-armed bandits with applications in maternal and child health. (2024). PRICAI 2024: Trends in artificial intelligence: 21st Pacific Rim International Conference on Artificial Intelligence, Kyoto, November 18-24. 15285, 165-178.
Available at: https://ink.library.smu.edu.sg/sis_research/9778
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1007/978-981-96-0128-8_15
Included in
Health Information Technology Commons, Operations Research, Systems Engineering and Industrial Engineering Commons, Theory and Algorithms Commons