File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Multi-Agent Reinforcement Learning in NOMA-Aided UAV Networks for Cellular Offloading

TitleMulti-Agent Reinforcement Learning in NOMA-Aided UAV Networks for Cellular Offloading
Authors
KeywordsDeep Q-network
non-orthogonal multiple access
reinforcement learning
unmanned aerial vehicle
Issue Date2022
Citation
IEEE Transactions on Wireless Communications, 2022, v. 21, n. 3, p. 1498-1512 How to Cite?
AbstractA novel framework is proposed for cellular offloading with the aid of multiple unmanned aerial vehicles (UAVs), while non-orthogonal multiple access (NOMA) technique is employed at each UAV to further improve the spectrum efficiency of the wireless network. The optimization problem of joint three-dimensional (3D) trajectory design and power allocation is formulated for maximizing the throughput. Since ground mobile users are considered as roaming continuously, the UAVs need to be re-deployed timely based on the movement of users. In an effort to solve this pertinent dynamic problem, a K-means based clustering algorithm is first adopted for periodically partitioning users. Afterward, a mutual deep Q-network (MDQN) algorithm is proposed to jointly determine the optimal 3D trajectory and power allocation of UAVs. In contrast to the conventional deep Q-network (DQN) algorithm, the MDQN algorithm enables the experience of multi-agent to be input into a shared neural network to shorten the training time with the assistance of state abstraction. Numerical results demonstrate that: 1) the proposed MDQN algorithm is capable of converging under minor constraints and has a faster convergence rate than the conventional DQN algorithm in the multi-agent case; 2) The achievable sum rate of the NOMA enhanced UAV network is 23% superior to the case of orthogonal multiple access (OMA); 3) By designing the optimal 3D trajectory of UAVs with the MDON algorithm, the sum rate of the network enjoys 142% and 56% gains than invoking the circular trajectory and the 2D trajectory, respectively.
Persistent Identifierhttp://hdl.handle.net/10722/349600
ISSN
2023 Impact Factor: 8.9
2023 SCImago Journal Rankings: 5.371

 

DC FieldValueLanguage
dc.contributor.authorZhong, Ruikang-
dc.contributor.authorLiu, Xiao-
dc.contributor.authorLiu, Yuanwei-
dc.contributor.authorChen, Yue-
dc.date.accessioned2024-10-17T06:59:37Z-
dc.date.available2024-10-17T06:59:37Z-
dc.date.issued2022-
dc.identifier.citationIEEE Transactions on Wireless Communications, 2022, v. 21, n. 3, p. 1498-1512-
dc.identifier.issn1536-1276-
dc.identifier.urihttp://hdl.handle.net/10722/349600-
dc.description.abstractA novel framework is proposed for cellular offloading with the aid of multiple unmanned aerial vehicles (UAVs), while non-orthogonal multiple access (NOMA) technique is employed at each UAV to further improve the spectrum efficiency of the wireless network. The optimization problem of joint three-dimensional (3D) trajectory design and power allocation is formulated for maximizing the throughput. Since ground mobile users are considered as roaming continuously, the UAVs need to be re-deployed timely based on the movement of users. In an effort to solve this pertinent dynamic problem, a K-means based clustering algorithm is first adopted for periodically partitioning users. Afterward, a mutual deep Q-network (MDQN) algorithm is proposed to jointly determine the optimal 3D trajectory and power allocation of UAVs. In contrast to the conventional deep Q-network (DQN) algorithm, the MDQN algorithm enables the experience of multi-agent to be input into a shared neural network to shorten the training time with the assistance of state abstraction. Numerical results demonstrate that: 1) the proposed MDQN algorithm is capable of converging under minor constraints and has a faster convergence rate than the conventional DQN algorithm in the multi-agent case; 2) The achievable sum rate of the NOMA enhanced UAV network is 23% superior to the case of orthogonal multiple access (OMA); 3) By designing the optimal 3D trajectory of UAVs with the MDON algorithm, the sum rate of the network enjoys 142% and 56% gains than invoking the circular trajectory and the 2D trajectory, respectively.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Wireless Communications-
dc.subjectDeep Q-network-
dc.subjectnon-orthogonal multiple access-
dc.subjectreinforcement learning-
dc.subjectunmanned aerial vehicle-
dc.titleMulti-Agent Reinforcement Learning in NOMA-Aided UAV Networks for Cellular Offloading-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TWC.2021.3104633-
dc.identifier.scopuseid_2-s2.0-85113298251-
dc.identifier.volume21-
dc.identifier.issue3-
dc.identifier.spage1498-
dc.identifier.epage1512-
dc.identifier.eissn1558-2248-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats