Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.
An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports / Huang, Yiwen; Huang, Wentao; Li, Ran; Huang, Tao; Li, Canbing; Tai, Nengling. - In: IEEE TRANSACTIONS ON SMART GRID. - ISSN 1949-3061. - 16:3(2025), pp. 2261-2277. [10.1109/TSG.2025.3547830]
An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports
Tao Huang;
2025
Abstract
Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.| File | Dimensione | Formato | |
|---|---|---|---|
|
An_Adaptive_MARL_Large_Model_for_Dispatch_Strategy_Generation_in_Logistics-Energy_Spatiotemporal_Coordination_of_Container_Seaports.pdf
accesso riservato
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
2.95 MB
Formato
Adobe PDF
|
2.95 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/3009065
