Publication Type
Journal Article
Version
acceptedVersion
Publication Date
1-2024
Abstract
Exploiting unmanned aerial vehicles (UAVs) to execute tasks is gaining growing popularity recently. To address the underlying task scheduling problem, conventional exact and heuristic algorithms encounter challenges such as rapidly increasing computation time and heavy reliance on domain knowledge, particularly when dealing with large-scale problems. The deep reinforcement learning (DRL) based methods that learn useful patterns from massive data demonstrate notable advantages. However, their decision space will become prohibitively huge as the problem scales up, thus deteriorating the computation efficiency. To alleviate this issue, we propose a double-level deep reinforcement learning (DL-DRL) approach based on a divide and conquer framework (DCF), where we decompose the task scheduling of multi-UAV into task allocation and route planning. Particularly, we design an encoder-decoder structured policy network in our upper-level DRL model to allocate the tasks to different UAVs, and we exploit another attention-based policy network in our lower-level DRL model to construct the route for each UAV, with the objective to maximize the total value of executed tasks given the maximum flight distance of the UAV. To effectively train the two models, we design an interactive training strategy (ITS), which includes pre-training, intensive training and alternate training. Experimental results show that our DL-DRL performs favorably against the learning-based and conventional baselines including the OR-Tools, in terms of solution quality and computation efficiency. We also verify the generalization performance of our approach by applying it to larger sizes of up to 1500 tasks and to different flight distances of UAVs. Moreover, we also show via an ablation study that our ITS can help achieve a balance between the performance and training efficiency. Our code is publicly available at https://faculty.csu.edu.cn/guohuawu/zh_CN/zdylm/193832/list/ index.htm.
Keywords
Deep reinforcement learning; divide and conquer-based framework; interactive training; multi-UAV task scheduling
Discipline
Operations Research, Systems Engineering and Industrial Engineering | Theory and Algorithms
Research Areas
Intelligent Systems and Optimization
Publication
IEEE Transactions on Automation Science and Engineering
First Page
1
Last Page
17
ISSN
1545-5955
Identifier
10.1109/TASE.2024.3358894
Publisher
Institute of Electrical and Electronics Engineers
Citation
MAO, Xiao; WU, Guohua; FAN, Mingfeng; CAO, Zhiguang; and PEDRYCZ, Witold.
DL-DRL: A double-level deep reinforcement learning approach for large-scale task scheduling of multi-UAV. (2024). IEEE Transactions on Automation Science and Engineering. 1-17.
Available at: https://ink.library.smu.edu.sg/sis_research/8698
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1109/TASE.2024.3358894
Included in
Operations Research, Systems Engineering and Industrial Engineering Commons, Theory and Algorithms Commons