Unmanned aerial base stations (UABSs) are a promising solution for improving coverage and capacity in vehicle-to-everything (V2X) communications, particularly in dense urban areas. However, their operation is constrained by onboard energy consumption, required for both flight and communication. Beamforming, while enhancing network performance, adds to this challenge due to its energy-intensive nature. This paper proposes a sequential and hierarchical decision-making framework for UABS operations, considering trajectory planning, dynamic beamforming, and radio resource assignment (RRA). While heuristic and optimal solutions are employed for trajectory planning and RRA respectively, the beamforming model is modeled as Markov decision process (MDP) to maximize served user demand, weighted by time-varying priorities, under strict energy constraints. Leveraging a dueling double deep q-network (3DQN) algorithm that penalizes energy budget violations, an agent policy for the beamformer is then trained. Simulation results demonstrate that the proposed approach outperforms static beamforming benchmarks and closely matches an ideal step-wise oracle, achieving a balance between energy efficiency and served user demand while adapting to dynamic V2X traffic conditions.
Efficient Dynamic Beamforming Activation for UAV-Enabled Vehicular Networks / Spampinato, L.; Amorosa, L. M.; Skocaj, M.; Vallero, G.; Renga, D.; Buratti, C.. - (2025), pp. 1-6. ( 2025 IEEE 36th International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC) Istanbul (Tur) 01-04 September 2025) [10.1109/PIMRC62392.2025.11274728].
Efficient Dynamic Beamforming Activation for UAV-Enabled Vehicular Networks
Renga D.;
2025
Abstract
Unmanned aerial base stations (UABSs) are a promising solution for improving coverage and capacity in vehicle-to-everything (V2X) communications, particularly in dense urban areas. However, their operation is constrained by onboard energy consumption, required for both flight and communication. Beamforming, while enhancing network performance, adds to this challenge due to its energy-intensive nature. This paper proposes a sequential and hierarchical decision-making framework for UABS operations, considering trajectory planning, dynamic beamforming, and radio resource assignment (RRA). While heuristic and optimal solutions are employed for trajectory planning and RRA respectively, the beamforming model is modeled as Markov decision process (MDP) to maximize served user demand, weighted by time-varying priorities, under strict energy constraints. Leveraging a dueling double deep q-network (3DQN) algorithm that penalizes energy budget violations, an agent policy for the beamformer is then trained. Simulation results demonstrate that the proposed approach outperforms static beamforming benchmarks and closely matches an ideal step-wise oracle, achieving a balance between energy efficiency and served user demand while adapting to dynamic V2X traffic conditions.| File | Dimensione | Formato | |
|---|---|---|---|
|
Dynamic_Beamforming_Activation_for_Green_UAV_Enabled_Vehicular_Networks (1).pdf
accesso aperto
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
Pubblico - Tutti i diritti riservati
Dimensione
1.9 MB
Formato
Adobe PDF
|
1.9 MB | Adobe PDF | Visualizza/Apri |
|
Efficient_Dynamic_Beamforming_Activation_for_UAV-Enabled_Vehicular_Networks.pdf
accesso riservato
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
2.6 MB
Formato
Adobe PDF
|
2.6 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/3010212
