We propose a Graph Transformer-based Reinforcement Learning Network Localization (GTRL) framework, which bridges graph representation learning with multi-agent decision making for cooperative Sensor Network Localization (SNL). Unlike conventional methods, GTRL casts localization as a cooperative policy learning problem, where a Graph Transformer (GT) encoder extracts spatial dependencies from graph-structured observations, and Multi-Agent Proximal Policy Optimization (MAPPO) learns per-node policies to refine positions. This paradigm shift enables agents to adaptively aggregate information and make distributed decisions, yielding robustness to noise, scalability to large networks, and efficiency under limited communication. Extensive experiments show that GTRL converges significantly faster than traditional methods, achieving higher accuracy with far fewer iterations. In addition, it generalizes well across diverse network layouts, anchor densities, communication ranges, and noise levels, outperforming state-of-the-art Graph Neural Network (GNN) localization methods.

Graph Neural Network-Based Reinforcement Learning for Cooperative Network Localization / Wu, Jinze; Li, Zhi; Lin, Zhiyun; Cheng, Hui; Zino, Lorenzo; Rizzo, Alessandro. - (2026), pp. 21111-21115. ( ICASSP 2026 - 2026 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) Barcelona (Spa) 03-08 May 2026) [10.1109/icassp55912.2026.11462755].

Graph Neural Network-Based Reinforcement Learning for Cooperative Network Localization

Wu, Jinze;Zino, Lorenzo;Rizzo, Alessandro
2026

Abstract

We propose a Graph Transformer-based Reinforcement Learning Network Localization (GTRL) framework, which bridges graph representation learning with multi-agent decision making for cooperative Sensor Network Localization (SNL). Unlike conventional methods, GTRL casts localization as a cooperative policy learning problem, where a Graph Transformer (GT) encoder extracts spatial dependencies from graph-structured observations, and Multi-Agent Proximal Policy Optimization (MAPPO) learns per-node policies to refine positions. This paradigm shift enables agents to adaptively aggregate information and make distributed decisions, yielding robustness to noise, scalability to large networks, and efficiency under limited communication. Extensive experiments show that GTRL converges significantly faster than traditional methods, achieving higher accuracy with far fewer iterations. In addition, it generalizes well across diverse network layouts, anchor densities, communication ranges, and noise levels, outperforming state-of-the-art Graph Neural Network (GNN) localization methods.
File in questo prodotto:
File Dimensione Formato  
ICASSP_2026.pdf

accesso riservato

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 2.06 MB
Formato Adobe PDF
2.06 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
2026_ICASSP_GNN_RL_Localization.pdf

accesso riservato

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 1.41 MB
Formato Adobe PDF
1.41 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/3010268