File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/GLOCOMW.2018.8644345
- Scopus: eid_2-s2.0-85063476546
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Deployment and Movement for Multiple Aerial Base Stations by Reinforcement Learning
Title | Deployment and Movement for Multiple Aerial Base Stations by Reinforcement Learning |
---|---|
Authors | |
Issue Date | 2018 |
Citation | 2018 IEEE Globecom Workshops, GC Wkshps 2018 - Proceedings, 2018, article no. 8644345 How to Cite? |
Abstract | A novel framework for Quality of experience (QoE)-driven deployment and movement of multiple unmanned aerial vehicles (UAVs) is proposed. The problem of joint non-concave 3D deployment and dynamic movement for maximizing the sum mean opinion score (MOS) of users is formulated, which is proved to be NP-hard. In an effort to solve this problem, we proposed a three-step approach to obtain 3D deployment and dynamic movement of multiple UAVs. More specifically, in the first step, GAK-means algorithm is invoked to obtain the cell partitioning of ground users. Secondly, Q-learning based deployment algorithm is proposed, in which each UAV is considered as an agent, making its own decision to obtain 3D position. In contrast to conventional genetic algorithm based learning algorithms, the proposed algorithm is capable of training the policy of making decision offline. Thirdly, Q-learning algorithm is invoked when the ground users roam. Unlike the other trajectory obtaining algorithms, the proposed approach enables each UAV learn its movement gradually through trials and errors, and updates the direction selection strategy until it reaches convergence. Numerical results reveal that the proposed 3D deployment scheme outperforms K-means algorithm and IGK algorithm with low complexity. Additionally, with the aid of proposed approach, 3D real-time dynamic movement of UAVs is obtained. |
Persistent Identifier | http://hdl.handle.net/10722/349315 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Liu, Xiao | - |
dc.contributor.author | Liu, Yuanwei | - |
dc.contributor.author | Chen, Yue | - |
dc.date.accessioned | 2024-10-17T06:57:43Z | - |
dc.date.available | 2024-10-17T06:57:43Z | - |
dc.date.issued | 2018 | - |
dc.identifier.citation | 2018 IEEE Globecom Workshops, GC Wkshps 2018 - Proceedings, 2018, article no. 8644345 | - |
dc.identifier.uri | http://hdl.handle.net/10722/349315 | - |
dc.description.abstract | A novel framework for Quality of experience (QoE)-driven deployment and movement of multiple unmanned aerial vehicles (UAVs) is proposed. The problem of joint non-concave 3D deployment and dynamic movement for maximizing the sum mean opinion score (MOS) of users is formulated, which is proved to be NP-hard. In an effort to solve this problem, we proposed a three-step approach to obtain 3D deployment and dynamic movement of multiple UAVs. More specifically, in the first step, GAK-means algorithm is invoked to obtain the cell partitioning of ground users. Secondly, Q-learning based deployment algorithm is proposed, in which each UAV is considered as an agent, making its own decision to obtain 3D position. In contrast to conventional genetic algorithm based learning algorithms, the proposed algorithm is capable of training the policy of making decision offline. Thirdly, Q-learning algorithm is invoked when the ground users roam. Unlike the other trajectory obtaining algorithms, the proposed approach enables each UAV learn its movement gradually through trials and errors, and updates the direction selection strategy until it reaches convergence. Numerical results reveal that the proposed 3D deployment scheme outperforms K-means algorithm and IGK algorithm with low complexity. Additionally, with the aid of proposed approach, 3D real-time dynamic movement of UAVs is obtained. | - |
dc.language | eng | - |
dc.relation.ispartof | 2018 IEEE Globecom Workshops, GC Wkshps 2018 - Proceedings | - |
dc.title | Deployment and Movement for Multiple Aerial Base Stations by Reinforcement Learning | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/GLOCOMW.2018.8644345 | - |
dc.identifier.scopus | eid_2-s2.0-85063476546 | - |
dc.identifier.spage | article no. 8644345 | - |
dc.identifier.epage | article no. 8644345 | - |