CrossFuNet: RGB and Depth Cross-Fusion Network for Hand Pose Estimation.
RGBD fusion
convolutional neural network
hand pose estimation
Journal
Sensors (Basel, Switzerland)
ISSN: 1424-8220
Titre abrégé: Sensors (Basel)
Pays: Switzerland
ID NLM: 101204366
Informations de publication
Date de publication:
11 Sep 2021
11 Sep 2021
Historique:
received:
02
08
2021
revised:
08
09
2021
accepted:
09
09
2021
entrez:
28
9
2021
pubmed:
29
9
2021
medline:
30
9
2021
Statut:
epublish
Résumé
Despite recent successes in hand pose estimation from RGB images or depth maps, inherent challenges remain. RGB-based methods suffer from heavy self-occlusions and depth ambiguity. Depth sensors rely heavily on distance and can only be used indoors, thus there are many limitations to the practical application of depth-based methods. The aforementioned challenges have inspired us to combine the two modalities to offset the shortcomings of the other. In this paper, we propose a novel RGB and depth information fusion network to improve the accuracy of 3D hand pose estimation, which is called CrossFuNet. Specifically, the RGB image and the paired depth map are input into two different subnetworks, respectively. The feature maps are fused in the fusion module in which we propose a completely new approach to combine the information from the two modalities. Then, the common method is used to regress the 3D key-points by heatmaps. We validate our model on two public datasets and the results reveal that our model outperforms the state-of-the-art methods.
Identifiants
pubmed: 34577302
pii: s21186095
doi: 10.3390/s21186095
pmc: PMC8473363
pii:
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM