Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking.

Siamese networks attentional mechanism convolutional neural network visual tracking

Journal

Sensors (Basel, Switzerland)
ISSN: 1424-8220
Titre abrégé: Sensors (Basel)
Pays: Switzerland
ID NLM: 101204366

Informations de publication

Date de publication:
20 Jul 2020
Historique:
received: 17 06 2020
revised: 03 07 2020
accepted: 15 07 2020
entrez: 24 7 2020
pubmed: 24 7 2020
medline: 24 7 2020
Statut: epublish

Résumé

We propose to improve the visual object tracking by introducing a soft mask based low-level feature fusion technique. The proposed technique is further strengthened by integrating channel and spatial attention mechanisms. The proposed approach is integrated within a Siamese framework to demonstrate its effectiveness for visual object tracking. The proposed soft mask is used to give more importance to the target regions as compared to the other regions to enable effective target feature representation and to increase discriminative power. The low-level feature fusion improves the tracker robustness against distractors. The channel attention is used to identify more discriminative channels for better target representation. The spatial attention complements the soft mask based approach to better localize the target objects in challenging tracking scenarios. We evaluated our proposed approach over five publicly available benchmark datasets and performed extensive comparisons with 39 state-of-the-art tracking algorithms. The proposed tracker demonstrates excellent performance compared to the existing state-of-the-art trackers.

Identifiants

pubmed: 32698339
pii: s20144021
doi: 10.3390/s20144021
pmc: PMC7412361
pii:
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Références

IEEE Trans Image Process. 2015 Dec;24(12):5630-44
pubmed: 26415202
Sensors (Basel). 2018 Dec 25;19(1):
pubmed: 30585239
IEEE Trans Pattern Anal Mach Intell. 2019 Nov;41(11):2770-2782
pubmed: 30106707
IEEE Trans Pattern Anal Mach Intell. 2015 Mar;37(3):583-96
pubmed: 26353263
IEEE Trans Image Process. 2019 Feb 21;:
pubmed: 30802856
IEEE Trans Cybern. 2020 Jul;50(7):3068-3080
pubmed: 31536029
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848
pubmed: 28463186
Sensors (Basel). 2020 Jul 06;20(13):
pubmed: 32640545
IEEE Trans Pattern Anal Mach Intell. 2017 Aug;39(8):1561-1575
pubmed: 27654137
IEEE Trans Pattern Anal Mach Intell. 2015 Sep;37(9):1834-48
pubmed: 26353130
IEEE Trans Pattern Anal Mach Intell. 2019 Jul 23;:
pubmed: 31352331
IEEE Trans Pattern Anal Mach Intell. 2019 Dec 04;:
pubmed: 31804928
IEEE Trans Cybern. 2018 Jun;48(6):1849-1861
pubmed: 28678728
IEEE Trans Pattern Anal Mach Intell. 2019 Jan;41(1):121-135
pubmed: 29990235

Auteurs

Mustansar Fiaz (M)

School of Computer Science and Engineering, Kyungpook National University, Daegu 41566, Korea.

Arif Mahmood (A)

Department of Computer Science, Information Technology University, Lahore 54000, Pakistan.

Soon Ki Jung (SK)

School of Computer Science and Engineering, Kyungpook National University, Daegu 41566, Korea.

Classifications MeSH