This paper introduces a novel physics-informed exploration strategy for a deep reinforcement learning (DRL)-based energy management system (EMS), specifically targeting the challenge of dealing with constrained action sets. RL-based controllers for electrified vehicle energy management systems have faced obstacles stemming from the selection of infeasible actions, obstructing their practical deployment. The absence of a mechanism for assessing control action feasibility prior to application has compounded this issue, primarily due to the model-free nature of RL-based controllers. Adding a safety layer to the RL-based controller addresses the abovementioned issue, but this often results in suboptimal policies and necessitates an in-depth understanding of the powertrain. Alternatively, theoretical remedies incorporate penalty terms into the immediate reward function to manage infeasible conditions. However, this approach can slow down the training process as the agent learns to avoid infeasible actions. To surmount these challenges, this paper introduces a novel physics-informed exploration strategy, coupled with prioritized experience replay, enabling the agent to swiftly learn to avoid selecting infeasible control actions without the need for a separate safety layer. Real-time simulation results highlight the superior performance of the proposed DRL-based controller over the baseline DRL-based controller with a safety layer, particularly in terms of overall fuel consumption.
Safe Reinforcement Learning for Energy Management of Electrified Vehicle with Novel Physics-Informed Exploration Strategy / Biswas, Atriya; Acquarone, Matteo; Wang, Hao; Miretti, Federico; Misul, Daniela Anna; Emadi, Ali. - In: IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION. - ISSN 2332-7782. - ELETTRONICO. - (2024). [10.1109/tte.2024.3361462]
Safe Reinforcement Learning for Energy Management of Electrified Vehicle with Novel Physics-Informed Exploration Strategy
Acquarone, Matteo;Miretti, Federico;Misul, Daniela Anna;
2024
Abstract
This paper introduces a novel physics-informed exploration strategy for a deep reinforcement learning (DRL)-based energy management system (EMS), specifically targeting the challenge of dealing with constrained action sets. RL-based controllers for electrified vehicle energy management systems have faced obstacles stemming from the selection of infeasible actions, obstructing their practical deployment. The absence of a mechanism for assessing control action feasibility prior to application has compounded this issue, primarily due to the model-free nature of RL-based controllers. Adding a safety layer to the RL-based controller addresses the abovementioned issue, but this often results in suboptimal policies and necessitates an in-depth understanding of the powertrain. Alternatively, theoretical remedies incorporate penalty terms into the immediate reward function to manage infeasible conditions. However, this approach can slow down the training process as the agent learns to avoid infeasible actions. To surmount these challenges, this paper introduces a novel physics-informed exploration strategy, coupled with prioritized experience replay, enabling the agent to swiftly learn to avoid selecting infeasible control actions without the need for a separate safety layer. Real-time simulation results highlight the superior performance of the proposed DRL-based controller over the baseline DRL-based controller with a safety layer, particularly in terms of overall fuel consumption.File | Dimensione | Formato | |
---|---|---|---|
Safe_Reinforcement_Learning_for_Energy_Management_of_Electrified_Vehicle_with_Novel_Physics-Informed_Exploration_Strategy.pdf
accesso aperto
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
PUBBLICO - Tutti i diritti riservati
Dimensione
2.2 MB
Formato
Adobe PDF
|
2.2 MB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2987319