Robot grasping has been widely studied in the last decade. Recently, Deep Learning made possible to achieve remarkable results in grasp pose estimation, using depth and RGB images. However, only few works consider the choice of the object to grasp. Moreover, they require a huge amount of data for generalizing to unseen object categories. For this reason, we introduce the Few-shot Semantic Grasping task where the objective is inferring a correct grasp given only five labelled images of a target unseen object. We propose a new deep learning architecture able to solve the aforementioned problem, leveraging on a Few-shot Semantic Segmentation module. We have evaluated the proposed model both in the Graspnet dataset and in a real scenario. In Graspnet, we achieve 40,95% accuracy in the Few-shot Semantic Grasping task, outperforming baseline approaches. In the real experiments, the results confirmed the generalization ability of the network.
FSG-Net: a deep learning model for semantic robot grasping through few-shot learning / Barcellona, Leonardo; Bacchin, Alberto; Gottardi, Alberto; Menegatti, Emanuele; Ghidoni, Stefano. - (2023), pp. 1793-1799. (Intervento presentato al convegno IEEE International Conference on Robotics and Automation (ICRA) tenutosi a Londra (UK) nel 29 May 2023 - 02 June 2023) [10.1109/ICRA48891.2023.10160618].
FSG-Net: a deep learning model for semantic robot grasping through few-shot learning
Barcellona, Leonardo;
2023
Abstract
Robot grasping has been widely studied in the last decade. Recently, Deep Learning made possible to achieve remarkable results in grasp pose estimation, using depth and RGB images. However, only few works consider the choice of the object to grasp. Moreover, they require a huge amount of data for generalizing to unseen object categories. For this reason, we introduce the Few-shot Semantic Grasping task where the objective is inferring a correct grasp given only five labelled images of a target unseen object. We propose a new deep learning architecture able to solve the aforementioned problem, leveraging on a Few-shot Semantic Segmentation module. We have evaluated the proposed model both in the Graspnet dataset and in a real scenario. In Graspnet, we achieve 40,95% accuracy in the Few-shot Semantic Grasping task, outperforming baseline approaches. In the real experiments, the results confirmed the generalization ability of the network.File | Dimensione | Formato | |
---|---|---|---|
ICRA2023_FSG.pdf
accesso aperto
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
PUBBLICO - Tutti i diritti riservati
Dimensione
5.27 MB
Formato
Adobe PDF
|
5.27 MB | Adobe PDF | Visualizza/Apri |
FSG-Net_a_Deep_Learning_model_for_Semantic_Robot_Grasping_through_Few-Shot_Learning.pdf
non disponibili
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
5.35 MB
Formato
Adobe PDF
|
5.35 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2977025