User association and power allocation for UAV-assisted networks: A distributed reinforcement learning approach

2020 ◽  
Vol 17 (12) ◽  
pp. 110-122
Author(s):  
Xin Guan ◽  
Yang Huang ◽  
Chao Dong ◽  
Qihui Wu
2021 ◽  
Vol 11 (9) ◽  
pp. 4135
Author(s):  
Chi-Kai Hsieh ◽  
Kun-Lin Chan ◽  
Feng-Tsun Chien

This paper studies the problem of joint power allocation and user association in wireless heterogeneous networks (HetNets) with a deep reinforcement learning (DRL)-based approach. This is a challenging problem since the action space is hybrid, consisting of continuous actions (power allocation) and discrete actions (device association). Instead of quantizing the continuous space (i.e., possible values of powers) into a set of discrete alternatives and applying traditional deep reinforcement approaches such as deep Q learning, we propose working on the hybrid space directly by using the novel parameterized deep Q-network (P-DQN) to update the learning policy and maximize the average cumulative reward. Furthermore, we incorporate the constraints of limited wireless backhaul capacity and the quality-of-service (QoS) of each user equipment (UE) into the learning process. Simulation results show that the proposed P-DQN outperforms the traditional approaches, such as the DQN and distance-based association, in terms of energy efficiency while satisfying the QoS and backhaul capacity constraints. The improvement in the energy efficiency of the proposed P-DQN on average may reach 77.6% and 140.6% over the traditional DQN and distance-based association approaches, respectively, in a HetNet with three SBS and five UEs.


Sign in / Sign up

Export Citation Format

Share Document