Publication Type
Journal Article
Version
acceptedVersion
Publication Date
6-2017
Abstract
Reinforcement learning is a plausible theoretical basis for developing self-learning, autonomous agents or robots that can effectively represent the world dynamics and efficiently learn the problem features to perform different tasks in different environments. The computational costs and complexities involved, however, are often prohibitive for real-world applications. This study introduces a scalable methodology to learn and transfer knowledge of the transition (and reward) models for model-based reinforcement learning in a complex world. We propose a variant formulation of Markov decision processes that supports efficient online-learning of the relevant problem features to approximate the world dynamics. We apply the new feature selection and dynamics approximation techniques in heterogeneous transfer learning, where the agent automatically maintains and adapts multiple representations of the world to cope with the different environments it encounters during its lifetime. We prove regret bounds for our approach, and empirically demonstrate its capability to quickly converge to a near optimal policy in both real and simulated environments.
Keywords
Model-based reinforcement learning, Online feature selection, Transfer learning
Discipline
Artificial Intelligence and Robotics
Research Areas
Intelligent Systems and Optimization
Publication
Artificial Intelligence
Volume
247
First Page
70
Last Page
94
ISSN
0004-3702
Identifier
10.1016/j.artint.2015.09.013
Publisher
Elsevier
Citation
Nguyen, Trung Thanh; Silander, Tomi; LI, Zhuoru; and Tze-Yun LEONG.
Scalable transfer learning in heterogeneous, dynamic environments. (2017). Artificial Intelligence. 247, 70-94.
Available at: https://ink.library.smu.edu.sg/sis_research/3039
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1016/j.artint.2015.09.013