This paper presents a comparison between twin- delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) reinforcement learning algorithms in the context of training robust navigation policies for Jackal robots. By leveraging an open-source framework and custom motion control environments, the study evaluates the performance, robustness, and transferability of the trained policies across a range of scenarios. The primary focus of the experiments is to assess the training process, the adaptability of the algorithms, and the robot’s ability to navigate in previously unseen environments. Moreover, the paper examines the influence of varying environ- ment complexities on the learning process and the generalization capabilities of the resulting policies. The results of this study aim to inform and guide the development of more efficient and practical reinforcement learning-based navigation policies for Jackal robots in real-world scenarios.
A Comparative Study of Twin Delayed Deep Deterministic Policy Gradient and Soft Actor-Critic Algorithms for Robot Exploration and Navigation in Unseen Environments / Ali, Romisaa. - In: WORLD ACADEMY OF SCIENCE, ENGINEERING AND TECHNOLOGY. - ISSN 2010-376X. - ELETTRONICO. - Vol:17:(In corso di stampa), pp. 1-7. (Intervento presentato al convegno Robot Motion Control tenutosi a Auckland, New Zealand nel December 04-05, 2023).
A Comparative Study of Twin Delayed Deep Deterministic Policy Gradient and Soft Actor-Critic Algorithms for Robot Exploration and Navigation in Unseen Environments
Romisaa Ali
In corso di stampa
Abstract
This paper presents a comparison between twin- delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) reinforcement learning algorithms in the context of training robust navigation policies for Jackal robots. By leveraging an open-source framework and custom motion control environments, the study evaluates the performance, robustness, and transferability of the trained policies across a range of scenarios. The primary focus of the experiments is to assess the training process, the adaptability of the algorithms, and the robot’s ability to navigate in previously unseen environments. Moreover, the paper examines the influence of varying environ- ment complexities on the learning process and the generalization capabilities of the resulting policies. The results of this study aim to inform and guide the development of more efficient and practical reinforcement learning-based navigation policies for Jackal robots in real-world scenarios.File | Dimensione | Formato | |
---|---|---|---|
25PT040005[5].pdf
accesso aperto
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
Creative commons
Dimensione
959.69 kB
Formato
Adobe PDF
|
959.69 kB | Adobe PDF | Visualizza/Apri |
paper_acceptance_letter.pdf
non disponibili
Descrizione: acceptence letter for oral presentation
Tipologia:
Altro materiale allegato
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
210.34 kB
Formato
Adobe PDF
|
210.34 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2982729