Publication Type
Conference Proceeding Article
Version
acceptedVersion
Publication Date
2-2024
Abstract
A popular framework for enforcing safe actions in Reinforcement Learning (RL) is Constrained RL, where trajectory based constraints on expected cost (or other cost measures) are employed to enforce safety and more importantly these constraints are enforced while maximizing expected reward. Most recent approaches for solving Constrained RL convert the trajectory based cost constraint into a surrogate problem that can be solved using minor modifications to RL methods. A key drawback with such approaches is an over or underestimation of the cost constraint at each state. Therefore, we provide an approach that does not modify the trajectory based cost constraint and instead imitates ``good'' trajectories and avoids ``bad'' trajectories generated from incrementally improving policies. We employ an oracle that utilizes a reward threshold (which is varied with learning) and the overall cost constraint to label trajectories as ``good'' or ``bad''. A key advantage of our approach is that we are able to work from any starting policy or set of trajectories and improve on it. In an exhaustive set of experiments, we demonstrate that our approach is able to outperform top benchmark approaches for solving Constrained RL problems, with respect to expected cost, CVaR cost, or even unknown cost constraints.
Discipline
Databases and Information Systems | Theory and Algorithms
Research Areas
Data Science and Engineering
Publication
Proceedings of the 38th Annual AAAI Conference on Artificial Intelligence, Vancouver, Canada, 2024 February 20-27
Identifier
10.48550/arXiv.2312.10385
Publisher
AAAI
City or Country
Washington
Citation
HOANG, Minh Huy; TIEN, Mai Anh; and VARAKANTHAM, Pradeep.
Imitate the good and avoid the bad: An incremental approach to safe reinforcement learning. (2024). Proceedings of the 38th Annual AAAI Conference on Artificial Intelligence, Vancouver, Canada, 2024 February 20-27.
Available at: https://ink.library.smu.edu.sg/sis_research/8594
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.48550/arXiv.2312.10385