Abstract
Unmanned aerial vehicles serving as aerial base stations (UAV-BSs) can be deployed to provide wireless connectivity to ground devices in events of increased network demand, points-of-failure in existing infrastructure, or disasters. However, it is challenging to conserve the energy of UAVs during prolonged coverage tasks, considering their limited on-board battery capacity. Reinforcement learning-based (RL) approaches have been previously used to improve energy utilization of multiple UAVs, however, a central cloud controller is assumed to have complete knowledge of the end-devices' locations, i.e., the controller periodically scans and sends updates for UAV decision-making. This assumption is impractical in dynamic network environments with UAVs serving mobile ground devices. To address this problem, we propose a decentralized Q-learning approach, where each UAVBS is equipped with an autonomous agent that maximizes the connectivity of mobile ground devices while improving its energy utilization. Experimental results show that the proposed design significantly outperforms the centralized approaches in jointly maximizing the number of connected ground devices and the energy utilization of the UAV-BSs.
| Original language | English |
|---|---|
| Pages (from-to) | 216-222 |
| Number of pages | 7 |
| Journal | Proceedings - IEEE Consumer Communications and Networking Conference, CCNC |
| DOIs | |
| Publication status | Published - 2022 |
| Externally published | Yes |
| Event | 19th IEEE Annual Consumer Communications and Networking Conference, CCNC 2022 - Virtual, Online, United States Duration: 8 Jan 2022 → 11 Jan 2022 |
Keywords
- energy management
- Reinforcement learning
- UAV base stations
- wireless connectivity
Fingerprint
Dive into the research topics of 'Energy-aware optimization of UAV base stations placement via decentralized multi-agent Q-learning'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver