Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.

An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports / Huang, Yiwen; Huang, Wentao; Li, Ran; Huang, Tao; Li, Canbing; Tai, Nengling. - In: IEEE TRANSACTIONS ON SMART GRID. - ISSN 1949-3061. - 16:3(2025), pp. 2261-2277. [10.1109/TSG.2025.3547830]

An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports

Tao Huang;
2025

Abstract

Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.
File in questo prodotto:
File Dimensione Formato  
An_Adaptive_MARL_Large_Model_for_Dispatch_Strategy_Generation_in_Logistics-Energy_Spatiotemporal_Coordination_of_Container_Seaports.pdf

accesso riservato

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 2.95 MB
Formato Adobe PDF
2.95 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/3009065