Nam H. Chu
Joint Speed Control and Energy Replenishment Optimization for UAV-Assisted IoT Data Collection With Deep Reinforcement Transfer Learning
Chu, Nam H.; Hoang, Dinh Thai; Nguyen, Diep N.; Van Huynh, Nguyen; Dutkiewicz, Eryk
Authors
Dinh Thai Hoang
Diep N. Nguyen
Nguyen Van Huynh
Eryk Dutkiewicz
Abstract
Unmanned-aerial-vehicle (UAV)-assisted data collection has been emerging as a prominent application due to its flexibility, mobility, and low operational cost. However, under the dynamic and uncertainty of Internet of Things data collection and energy replenishment processes, optimizing the performance for UAV collectors is a very challenging task. Thus, this article introduces a novel framework that jointly optimizes the flying speed and energy replenishment for each UAV to significantly improve the overall system performance (e.g., data collection and energy usage efficiency). Specifically, we first develop a Markov decision process to help the UAV automatically and dynamically make optimal decisions under the dynamics and uncertainties of the environment. Although traditional reinforcement learning algorithms, such as Q-learning and deep Q-learning, can help the UAV to obtain the optimal policy, they often take a long time to converge and require high computational complexity. Therefore, it is impractical to deploy these conventional methods on UAVs with limited computing capacity and energy resource. To that end, we develop advanced transfer learning techniques that allow UAVs to “share” and “transfer” learning knowledge, thereby reducing the learning time as well as significantly improving learning quality. Extensive simulations demonstrate that our proposed solution can improve the average data collection performance of the system up to 200% and reduce the convergence time up to 50% compared with those of conventional methods.
Citation
Chu, N. H., Hoang, D. T., Nguyen, D. N., Van Huynh, N., & Dutkiewicz, E. (2023). Joint Speed Control and Energy Replenishment Optimization for UAV-Assisted IoT Data Collection With Deep Reinforcement Transfer Learning. IEEE Internet of Things, 10(7), 5778-5793. https://doi.org/10.1109/jiot.2022.3151201
Journal Article Type | Article |
---|---|
Acceptance Date | Feb 7, 2022 |
Online Publication Date | Feb 14, 2022 |
Publication Date | 2023-04 |
Deposit Date | Mar 29, 2023 |
Print ISSN | 2327-4662 |
Electronic ISSN | 2327-4662 |
Publisher | Institute of Electrical and Electronics Engineers |
Peer Reviewed | Peer Reviewed |
Volume | 10 |
Issue | 7 |
Pages | 5778-5793 |
DOI | https://doi.org/10.1109/jiot.2022.3151201 |
Keywords | Deep reinforcement learning (DRL), Internet of Things (IoT) data collection, Markov decision process (MDP), transfer learning (TL), unmanned aerial vehicle (UAV) |
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search