File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/ICCWorkshops49005.2020.9145457
- Scopus: eid_2-s2.0-85090280958
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Distributed reinforcement learning for NOMA-enabled mobile edge computing
Title | Distributed reinforcement learning for NOMA-enabled mobile edge computing |
---|---|
Authors | |
Issue Date | 2020 |
Citation | 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020 - Proceedings, 2020, article no. 9145457 How to Cite? |
Abstract | A novel non-orthogonal multiple access (NOMA) enabled cache-aided mobile edge computing (MEC) framework is proposed, for minimizing the sum energy consumption. The NOMA strategy enables mobile users to offload computation tasks to the access point (AP) simultaneously, which improves the spectrum efficiency. In this article, the considered resource allocation problem is formulated as a long-term reward maximization problem that involves a joint optimization of task offloading decision, computation resource allocation, and caching decision. To tackle this nontrivial problem, a single-agent Q-learning (SAQ-learning) algorithm is invoked to learn a long-term resource allocation strategy from historical experience. Moreover, a Bayesian learning automata (BLA) based multi-agent Q-learning (MAQ-learning) algorithm is proposed for task offloading decisions. More specifically, a BLA based action select scheme is proposed for the agents in MAQ-learning to select the optimal actions in every state. The proposed BLA based action selection scheme is instantaneously self-correcting, consequently, if the probabilities of two computing models (i.e., local computing and offloading computing) are not equal, the optimal action unveils eventually. Extensive simulations demonstrate that: 1) The proposed cache-aided NOMA MEC framework significantly outperforms the other representative benchmark schemes under various network setups. 2) The effectiveness of the proposed BAL-MAQ-learning algorithm is confirmed from the comparison with the results of conventional reinforcement learning algorithms. |
Persistent Identifier | http://hdl.handle.net/10722/349464 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yang, Zhong | - |
dc.contributor.author | Liu, Yuanwei | - |
dc.contributor.author | Chen, Yue | - |
dc.date.accessioned | 2024-10-17T06:58:42Z | - |
dc.date.available | 2024-10-17T06:58:42Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020 - Proceedings, 2020, article no. 9145457 | - |
dc.identifier.uri | http://hdl.handle.net/10722/349464 | - |
dc.description.abstract | A novel non-orthogonal multiple access (NOMA) enabled cache-aided mobile edge computing (MEC) framework is proposed, for minimizing the sum energy consumption. The NOMA strategy enables mobile users to offload computation tasks to the access point (AP) simultaneously, which improves the spectrum efficiency. In this article, the considered resource allocation problem is formulated as a long-term reward maximization problem that involves a joint optimization of task offloading decision, computation resource allocation, and caching decision. To tackle this nontrivial problem, a single-agent Q-learning (SAQ-learning) algorithm is invoked to learn a long-term resource allocation strategy from historical experience. Moreover, a Bayesian learning automata (BLA) based multi-agent Q-learning (MAQ-learning) algorithm is proposed for task offloading decisions. More specifically, a BLA based action select scheme is proposed for the agents in MAQ-learning to select the optimal actions in every state. The proposed BLA based action selection scheme is instantaneously self-correcting, consequently, if the probabilities of two computing models (i.e., local computing and offloading computing) are not equal, the optimal action unveils eventually. Extensive simulations demonstrate that: 1) The proposed cache-aided NOMA MEC framework significantly outperforms the other representative benchmark schemes under various network setups. 2) The effectiveness of the proposed BAL-MAQ-learning algorithm is confirmed from the comparison with the results of conventional reinforcement learning algorithms. | - |
dc.language | eng | - |
dc.relation.ispartof | 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020 - Proceedings | - |
dc.title | Distributed reinforcement learning for NOMA-enabled mobile edge computing | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/ICCWorkshops49005.2020.9145457 | - |
dc.identifier.scopus | eid_2-s2.0-85090280958 | - |
dc.identifier.spage | article no. 9145457 | - |
dc.identifier.epage | article no. 9145457 | - |