Deep Q-Learning based Resource Management in UAV-assisted Wireless Powered IoT Networks

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
ICC 2020 - 2020 IEEE International Conference on Communications (ICC), 2020, 2020-June, pp. 1-6
Issue Date:
2020-07-27
Filename Description Size
09149282.pdfPublished version241.12 kB
Adobe PDF
Full metadata record
In Unmanned Aerial Vehicle (UAV)-assisted Wireless Powered Internet of Things (IoT), the UAV is employed to charge the IoT nodes remotely via Wireless Power Transfer (WPT) and collect their data. A key challenge of resource management for WPT and data collection is preventing battery drainage and butter overflow of the ground IoT nodes in the presence of highly dynamic airborne channels. In this paper, we consider the resource management problem in practical scenarios, where the UAV has no a-prior information on battery levels and data queue lengths of the nodes. We formulate the resource management of UAV-assisted WPT and data collection as Markov Decision Process (MDP), where the states consist of battery levels and data queue lengths of the IoT nodes, channel qualities, and positions of the UAV. A deep Q-learning based resource management is proposed to minimize the overall data packet loss of the IoT nodes, by optimally deciding the IoT node for data collection and power transfer, and the associated modulation scheme of the IoT node.
Please use this identifier to cite or link to this item: