Deep Reinforcement Learning based Energy-Efficient Multi-UAV Data Collection for IoT Networks
dc.contributor.advisor | Wang, Ping | |
dc.contributor.advisor | Nguyen, Uyen Trang | |
dc.contributor.author | Khodaparast, Seyed Saeed | |
dc.date.accessioned | 2022-03-03T13:57:59Z | |
dc.date.available | 2022-03-03T13:57:59Z | |
dc.date.copyright | 2021-09 | |
dc.date.issued | 2022-03-03 | |
dc.date.updated | 2022-03-03T13:57:59Z | |
dc.degree.discipline | Electrical and Computer Engineering | |
dc.degree.level | Master's | |
dc.degree.name | MASc - Master of Applied Science | |
dc.description.abstract | Unmanned aerial vehicles (UAVs) are regarded as an emerging technology, which can be effectively utilized to perform the data collection tasks in the Internet of Things (IoT) networks. However, both the UAVs and the sensors in these networks are energy-limited devices, which necessitates an energy-efficient data collection procedure to ensure the network lifetime. In this thesis, we propose a multi-UAV-assisted network, where the UAVs fly to the ground sensors and control the sensor's transmit power during the data collection time. Our goal is to minimize the total energy consumption of the UAVs and the sensors, which is needed to accomplish the data collection mission. We formulate this problem into three sub-problems of single UAV navigation, sensor power control, and multi-UAV scheduling, and model each part as a finite-horizon Markov Decision Process (MDP). We deploy deep reinforcement learning (DRL)-based frameworks to solve each part. Specifically, we use the deep deterministic policy gradient (DDPG) method to generate the best trajectory for the UAVs in an obstacle-constrained environment, given its starting position and the target sensor. We also deploy DDPG to control the sensor's transmit power during data collection. To schedule activity plans for each UAV to visit the sensors, we propose a multi-agent deep Q-learning (DQL) approach by taking the energy consumption of the UAVs on each path into account. Our simulations show that the UAVs can find a safe and optimal path for each of their trips. Continuous power control of the sensors achieves better performance than the fixed power and fixed rate approaches in terms of the sensor's energy consumption and the data collection completion time. In addition, compared to the two commonly used baselines, our scheduling framework achieves better and near-optimal results in the simulated scenario. | |
dc.identifier.uri | http://hdl.handle.net/10315/39061 | |
dc.language | en | |
dc.rights | Author owns copyright, except where explicitly noted. Please contact the author directly with licensing requests. | |
dc.subject | Engineering | |
dc.subject.keywords | Data collection | |
dc.subject.keywords | Unmanned aerial vehicle (UAV) | |
dc.subject.keywords | Internet of Things (IoT) | |
dc.subject.keywords | Deep reinforcement learning (DRL) | |
dc.subject.keywords | Energy consumption | |
dc.title | Deep Reinforcement Learning based Energy-Efficient Multi-UAV Data Collection for IoT Networks | |
dc.type | Electronic Thesis or Dissertation |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Khodaparast_SeyedSaeed_2021_Masters.pdf
- Size:
- 1.68 MB
- Format:
- Adobe Portable Document Format
- Description: