Transparency-Aware Segmentation of Glass Objects to Train RGB-Based Pose Estimators.
bounding box
neural networks
pose estimation
segmentation
training data
transparent objects
Journal
Sensors (Basel, Switzerland)
ISSN: 1424-8220
Titre abrégé: Sensors (Basel)
Pays: Switzerland
ID NLM: 101204366
Informations de publication
Date de publication:
10 Jan 2024
10 Jan 2024
Historique:
received:
04
12
2023
revised:
22
12
2023
accepted:
25
12
2023
medline:
23
1
2024
pubmed:
23
1
2024
entrez:
23
1
2024
Statut:
epublish
Résumé
Robotic manipulation requires object pose knowledge for the objects of interest. In order to perform typical household chores, a robot needs to be able to estimate 6D poses for objects such as water glasses or salad bowls. This is especially difficult for glass objects, as for these, depth data are mostly disturbed, and in RGB images, occluded objects are still visible. Thus, in this paper, we propose to redefine the ground-truth for training RGB-based pose estimators in two ways: (a) we apply a transparency-aware multisegmentation, in which an image pixel can belong to more than one object, and (b) we use transparency-aware bounding boxes, which always enclose whole objects, even if parts of an object are formally occluded by another object. The latter approach ensures that the size and scale of an object remain more consistent across different images. We train our pose estimator, which was originally designed for opaque objects, with three different ground-truth types on the ClearPose dataset. Just by changing the training data to our transparency-aware segmentation, with no additional glass-specific feature changes in the estimator, the ADD-S AUC value increases by 4.3%. Such a multisegmentation can be created for every dataset that provides a 3D model of the object and its ground-truth pose.
Identifiants
pubmed: 38257525
pii: s24020432
doi: 10.3390/s24020432
pii:
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Subventions
Organisme : Deutsche Forschungsgemeinschaft
ID : CRC 1320