This paper presents a comparison between twin- delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) reinforcement learning algorithms in the context of training robust navigation policies for Jackal robots. By leveraging an open-source framework and custom motion control environments, the study evaluates the performance, robustness, and transferability of the trained policies across a range of scenarios. The primary focus of the experiments is to assess the training process, the adaptability of the algorithms, and the robot’s ability to navigate in previously unseen environments. Moreover, the paper examines the influence of varying environ- ment complexities on the learning process and the generalization capabilities of the resulting policies. The results of this study aim to inform and guide the development of more efficient and practical reinforcement learning-based navigation policies for Jackal robots in real-world scenarios.

A Comparative Study of Twin Delayed Deep Deterministic Policy Gradient and Soft Actor-Critic Algorithms for Robot Exploration and Navigation in Unseen Environments / Ali, Romisaa. - In: WORLD ACADEMY OF SCIENCE, ENGINEERING AND TECHNOLOGY. - ISSN 2010-376X. - ELETTRONICO. - Vol:17:(In corso di stampa), pp. 1-7. (Intervento presentato al convegno Robot Motion Control tenutosi a Auckland, New Zealand nel December 04-05, 2023).

A Comparative Study of Twin Delayed Deep Deterministic Policy Gradient and Soft Actor-Critic Algorithms for Robot Exploration and Navigation in Unseen Environments

Romisaa Ali
In corso di stampa

Abstract

This paper presents a comparison between twin- delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) reinforcement learning algorithms in the context of training robust navigation policies for Jackal robots. By leveraging an open-source framework and custom motion control environments, the study evaluates the performance, robustness, and transferability of the trained policies across a range of scenarios. The primary focus of the experiments is to assess the training process, the adaptability of the algorithms, and the robot’s ability to navigate in previously unseen environments. Moreover, the paper examines the influence of varying environ- ment complexities on the learning process and the generalization capabilities of the resulting policies. The results of this study aim to inform and guide the development of more efficient and practical reinforcement learning-based navigation policies for Jackal robots in real-world scenarios.
File in questo prodotto:
File Dimensione Formato  
25PT040005[5].pdf

accesso aperto

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: Creative commons
Dimensione 959.69 kB
Formato Adobe PDF
959.69 kB Adobe PDF Visualizza/Apri
paper_acceptance_letter.pdf

non disponibili

Descrizione: acceptence letter for oral presentation
Tipologia: Altro materiale allegato
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 210.34 kB
Formato Adobe PDF
210.34 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2982729