Robot grasping has been widely studied in the last decade. Recently, Deep Learning made possible to achieve remarkable results in grasp pose estimation, using depth and RGB images. However, only few works consider the choice of the object to grasp. Moreover, they require a huge amount of data for generalizing to unseen object categories. For this reason, we introduce the Few-shot Semantic Grasping task where the objective is inferring a correct grasp given only five labelled images of a target unseen object. We propose a new deep learning architecture able to solve the aforementioned problem, leveraging on a Few-shot Semantic Segmentation module. We have evaluated the proposed model both in the Graspnet dataset and in a real scenario. In Graspnet, we achieve 40,95% accuracy in the Few-shot Semantic Grasping task, outperforming baseline approaches. In the real experiments, the results confirmed the generalization ability of the network.

FSG-Net: a deep learning model for semantic robot grasping through few-shot learning / Barcellona, Leonardo; Bacchin, Alberto; Gottardi, Alberto; Menegatti, Emanuele; Ghidoni, Stefano. - (2023), pp. 1793-1799. (Intervento presentato al convegno IEEE International Conference on Robotics and Automation (ICRA) tenutosi a Londra (UK) nel 29 May 2023 - 02 June 2023) [10.1109/ICRA48891.2023.10160618].

FSG-Net: a deep learning model for semantic robot grasping through few-shot learning

Barcellona, Leonardo;
2023

Abstract

Robot grasping has been widely studied in the last decade. Recently, Deep Learning made possible to achieve remarkable results in grasp pose estimation, using depth and RGB images. However, only few works consider the choice of the object to grasp. Moreover, they require a huge amount of data for generalizing to unseen object categories. For this reason, we introduce the Few-shot Semantic Grasping task where the objective is inferring a correct grasp given only five labelled images of a target unseen object. We propose a new deep learning architecture able to solve the aforementioned problem, leveraging on a Few-shot Semantic Segmentation module. We have evaluated the proposed model both in the Graspnet dataset and in a real scenario. In Graspnet, we achieve 40,95% accuracy in the Few-shot Semantic Grasping task, outperforming baseline approaches. In the real experiments, the results confirmed the generalization ability of the network.
2023
979-8-3503-2365-8
File in questo prodotto:
File Dimensione Formato  
ICRA2023_FSG.pdf

accesso aperto

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: PUBBLICO - Tutti i diritti riservati
Dimensione 5.27 MB
Formato Adobe PDF
5.27 MB Adobe PDF Visualizza/Apri
FSG-Net_a_Deep_Learning_model_for_Semantic_Robot_Grasping_through_Few-Shot_Learning.pdf

non disponibili

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 5.35 MB
Formato Adobe PDF
5.35 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2977025