File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Multi-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks

TitleMulti-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks
Authors
KeywordsDynamic resource allocation
multi-agent reinforcement learning (MARL)
stochastic games
UAV communications
Issue Date2020
Citation
IEEE Transactions on Wireless Communications, 2020, v. 19, n. 2, p. 729-743 How to Cite?
AbstractUnmanned aerial vehicles (UAVs) are capable of serving as aerial base stations (BSs) for providing both cost-effective and on-demand wireless communications. This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards. More particularly, each UAV communicates with a ground user by automatically selecting its communicating user, power level and subchannel without any information exchange among UAVs. To model the dynamics and uncertainty in environments, we formulate the long-term resource allocation problem as a stochastic game for maximizing the expected rewards, where each UAV becomes a learning agent and each resource allocation solution corresponds to an action taken by the UAVs. Afterwards, we develop a multi-agent reinforcement learning (MARL) framework that each agent discovers its best strategy according to its local observations using learning. More specifically, we propose an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning. Finally, simulation results reveal that: 1) appropriate parameters for exploitation and exploration are capable of enhancing the performance of the proposed MARL based resource allocation algorithm; 2) the proposed MARL algorithm provides acceptable performance compared to the case with complete information exchanges among UAVs. By doing so, it strikes a good tradeoff between performance gains and information exchange overheads.
Persistent Identifierhttp://hdl.handle.net/10722/349404
ISSN
2023 Impact Factor: 8.9
2023 SCImago Journal Rankings: 5.371

 

DC FieldValueLanguage
dc.contributor.authorCui, Jingjing-
dc.contributor.authorLiu, Yuanwei-
dc.contributor.authorNallanathan, Arumugam-
dc.date.accessioned2024-10-17T06:58:18Z-
dc.date.available2024-10-17T06:58:18Z-
dc.date.issued2020-
dc.identifier.citationIEEE Transactions on Wireless Communications, 2020, v. 19, n. 2, p. 729-743-
dc.identifier.issn1536-1276-
dc.identifier.urihttp://hdl.handle.net/10722/349404-
dc.description.abstractUnmanned aerial vehicles (UAVs) are capable of serving as aerial base stations (BSs) for providing both cost-effective and on-demand wireless communications. This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards. More particularly, each UAV communicates with a ground user by automatically selecting its communicating user, power level and subchannel without any information exchange among UAVs. To model the dynamics and uncertainty in environments, we formulate the long-term resource allocation problem as a stochastic game for maximizing the expected rewards, where each UAV becomes a learning agent and each resource allocation solution corresponds to an action taken by the UAVs. Afterwards, we develop a multi-agent reinforcement learning (MARL) framework that each agent discovers its best strategy according to its local observations using learning. More specifically, we propose an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning. Finally, simulation results reveal that: 1) appropriate parameters for exploitation and exploration are capable of enhancing the performance of the proposed MARL based resource allocation algorithm; 2) the proposed MARL algorithm provides acceptable performance compared to the case with complete information exchanges among UAVs. By doing so, it strikes a good tradeoff between performance gains and information exchange overheads.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Wireless Communications-
dc.subjectDynamic resource allocation-
dc.subjectmulti-agent reinforcement learning (MARL)-
dc.subjectstochastic games-
dc.subjectUAV communications-
dc.titleMulti-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TWC.2019.2935201-
dc.identifier.scopuseid_2-s2.0-85079784199-
dc.identifier.volume19-
dc.identifier.issue2-
dc.identifier.spage729-
dc.identifier.epage743-
dc.identifier.eissn1558-2248-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats