We propose a Graph Transformer-based Reinforcement Learning Network Localization (GTRL) framework, which bridges graph representation learning with multi-agent decision making for cooperative Sensor Network Localization (SNL). Unlike conventional methods, GTRL casts localization as a cooperative policy learning problem, where a Graph Transformer (GT) encoder extracts spatial dependencies from graph-structured observations, and Multi-Agent Proximal Policy Optimization (MAPPO) learns per-node policies to refine positions. This paradigm shift enables agents to adaptively aggregate information and make distributed decisions, yielding robustness to noise, scalability to large networks, and efficiency under limited communication. Extensive experiments show that GTRL converges significantly faster than traditional methods, achieving higher accuracy with far fewer iterations. In addition, it generalizes well across diverse network layouts, anchor densities, communication ranges, and noise levels, outperforming state-of-the-art Graph Neural Network (GNN) localization methods.
Graph Neural Network-Based Reinforcement Learning for Cooperative Network Localization / Wu, Jinze; Li, Zhi; Lin, Zhiyun; Cheng, Hui; Zino, Lorenzo; Rizzo, Alessandro. - (2026), pp. 21111-21115. ( ICASSP 2026 - 2026 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) Barcelona (Spa) 03-08 May 2026) [10.1109/icassp55912.2026.11462755].
Graph Neural Network-Based Reinforcement Learning for Cooperative Network Localization
Wu, Jinze;Zino, Lorenzo;Rizzo, Alessandro
2026
Abstract
We propose a Graph Transformer-based Reinforcement Learning Network Localization (GTRL) framework, which bridges graph representation learning with multi-agent decision making for cooperative Sensor Network Localization (SNL). Unlike conventional methods, GTRL casts localization as a cooperative policy learning problem, where a Graph Transformer (GT) encoder extracts spatial dependencies from graph-structured observations, and Multi-Agent Proximal Policy Optimization (MAPPO) learns per-node policies to refine positions. This paradigm shift enables agents to adaptively aggregate information and make distributed decisions, yielding robustness to noise, scalability to large networks, and efficiency under limited communication. Extensive experiments show that GTRL converges significantly faster than traditional methods, achieving higher accuracy with far fewer iterations. In addition, it generalizes well across diverse network layouts, anchor densities, communication ranges, and noise levels, outperforming state-of-the-art Graph Neural Network (GNN) localization methods.| File | Dimensione | Formato | |
|---|---|---|---|
|
ICASSP_2026.pdf
accesso riservato
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
2.06 MB
Formato
Adobe PDF
|
2.06 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
|
2026_ICASSP_GNN_RL_Localization.pdf
accesso riservato
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
1.41 MB
Formato
Adobe PDF
|
1.41 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/3010268
