Generation of Tactile Data From 3D Vision and Target Robotic Grasps.
Journal
IEEE transactions on haptics
ISSN: 2329-4051
Titre abrégé: IEEE Trans Haptics
Pays: United States
ID NLM: 101491191
Informations de publication
Date de publication:
Historique:
pubmed:
4
8
2020
medline:
26
10
2021
entrez:
4
8
2020
Statut:
ppublish
Résumé
Tactile perception is a rich source of information for robotic grasping: it allows a robot to identify a grasped object and assess the stability of a grasp, among other things. However, the tactile sensor must come into contact with the target object in order to produce readings. As a result, tactile data can only be attained if a real contact is made. We propose to overcome this restriction by employing a method that models the behaviour of a tactile sensor using 3D vision and grasp information as a stimulus. Our system regresses the quantified tactile response that would be experienced if this grasp were performed on the object. We experiment with 16 items and 4 tactile data modalities to show that our proposal learns this task with low error.
Identifiants
pubmed: 32746383
doi: 10.1109/TOH.2020.3011899
doi:
Types de publication
Journal Article
Research Support, Non-U.S. Gov't
Langues
eng
Sous-ensembles de citation
IM