This paper introduces a novel physics-informed exploration strategy for a deep reinforcement learning (DRL)-based energy management system (EMS), specifically targeting the challenge of dealing with constrained action sets. RL-based controllers for electrified vehicle energy management systems have faced obstacles stemming from the selection of infeasible actions, obstructing their practical deployment. The absence of a mechanism for assessing control action feasibility prior to application has compounded this issue, primarily due to the model-free nature of RL-based controllers. Adding a safety layer to the RL-based controller addresses the abovementioned issue, but this often results in suboptimal policies and necessitates an in-depth understanding of the powertrain. Alternatively, theoretical remedies incorporate penalty terms into the immediate reward function to manage infeasible conditions. However, this approach can slow down the training process as the agent learns to avoid infeasible actions. To surmount these challenges, this paper introduces a novel physics-informed exploration strategy, coupled with prioritized experience replay, enabling the agent to swiftly learn to avoid selecting infeasible control actions without the need for a separate safety layer. Real-time simulation results highlight the superior performance of the proposed DRL-based controller over the baseline DRL-based controller with a safety layer, particularly in terms of overall fuel consumption.

Safe Reinforcement Learning for Energy Management of Electrified Vehicle with Novel Physics-Informed Exploration Strategy / Biswas, Atriya; Acquarone, Matteo; Wang, Hao; Miretti, Federico; Misul, Daniela Anna; Emadi, Ali. - In: IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION. - ISSN 2332-7782. - ELETTRONICO. - (2024). [10.1109/tte.2024.3361462]

Safe Reinforcement Learning for Energy Management of Electrified Vehicle with Novel Physics-Informed Exploration Strategy

Acquarone, Matteo;Miretti, Federico;Misul, Daniela Anna;
2024

Abstract

This paper introduces a novel physics-informed exploration strategy for a deep reinforcement learning (DRL)-based energy management system (EMS), specifically targeting the challenge of dealing with constrained action sets. RL-based controllers for electrified vehicle energy management systems have faced obstacles stemming from the selection of infeasible actions, obstructing their practical deployment. The absence of a mechanism for assessing control action feasibility prior to application has compounded this issue, primarily due to the model-free nature of RL-based controllers. Adding a safety layer to the RL-based controller addresses the abovementioned issue, but this often results in suboptimal policies and necessitates an in-depth understanding of the powertrain. Alternatively, theoretical remedies incorporate penalty terms into the immediate reward function to manage infeasible conditions. However, this approach can slow down the training process as the agent learns to avoid infeasible actions. To surmount these challenges, this paper introduces a novel physics-informed exploration strategy, coupled with prioritized experience replay, enabling the agent to swiftly learn to avoid selecting infeasible control actions without the need for a separate safety layer. Real-time simulation results highlight the superior performance of the proposed DRL-based controller over the baseline DRL-based controller with a safety layer, particularly in terms of overall fuel consumption.
File in questo prodotto:
File Dimensione Formato  
Safe_Reinforcement_Learning_for_Energy_Management_of_Electrified_Vehicle_with_Novel_Physics-Informed_Exploration_Strategy.pdf

non disponibili

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 2.2 MB
Formato Adobe PDF
2.2 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2987319
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo