Deep Learning-Based Multimodal Data Fusion: Case Study in Food Intake Episodes Detection Using Wearable Sensors.
covariance distribution
data fusion
deep learning
food intake episode
image processing
wearable sensors
Journal
JMIR mHealth and uHealth
ISSN: 2291-5222
Titre abrégé: JMIR Mhealth Uhealth
Pays: Canada
ID NLM: 101624439
Informations de publication
Date de publication:
28 01 2021
28 01 2021
Historique:
received:
29
06
2020
accepted:
18
12
2020
revised:
10
11
2020
entrez:
28
1
2021
pubmed:
29
1
2021
medline:
9
3
2021
Statut:
epublish
Résumé
Multimodal wearable technologies have brought forward wide possibilities in human activity recognition, and more specifically personalized monitoring of eating habits. The emerging challenge now is the selection of most discriminative information from high-dimensional data collected from multiple sources. The available fusion algorithms with their complex structure are poorly adopted to the computationally constrained environment which requires integrating information directly at the source. As a result, more simple low-level fusion methods are needed. In the absence of a data combining process, the cost of directly applying high-dimensional raw data to a deep classifier would be computationally expensive with regard to the response time, energy consumption, and memory requirement. Taking this into account, we aimed to develop a data fusion technique in a computationally efficient way to achieve a more comprehensive insight of human activity dynamics in a lower dimension. The major objective was considering statistical dependency of multisensory data and exploring intermodality correlation patterns for different activities. In this technique, the information in time (regardless of the number of sources) is transformed into a 2D space that facilitates classification of eating episodes from others. This is based on a hypothesis that data captured by various sensors are statistically associated with each other and the covariance matrix of all these signals has a unique distribution correlated with each activity which can be encoded on a contour representation. These representations are then used as input of a deep model to learn specific patterns associated with specific activity. In order to show the generalizability of the proposed fusion algorithm, 2 different scenarios were taken into account. These scenarios were different in terms of temporal segment size, type of activity, wearable device, subjects, and deep learning architecture. The first scenario used a data set in which a single participant performed a limited number of activities while wearing the Empatica E4 wristband. In the second scenario, a data set related to the activities of daily living was used where 10 different participants wore inertial measurement units while performing a more complex set of activities. The precision metric obtained from leave-one-subject-out cross-validation for the second scenario reached 0.803. The impact of missing data on performance degradation was also evaluated. To conclude, the proposed fusion technique provides the possibility of embedding joint variability information over different modalities in just a single 2D representation which results in obtaining a more global view of different aspects of daily human activities at hand, and yet preserving the desired performance level in activity recognition.
Sections du résumé
BACKGROUND
Multimodal wearable technologies have brought forward wide possibilities in human activity recognition, and more specifically personalized monitoring of eating habits. The emerging challenge now is the selection of most discriminative information from high-dimensional data collected from multiple sources. The available fusion algorithms with their complex structure are poorly adopted to the computationally constrained environment which requires integrating information directly at the source. As a result, more simple low-level fusion methods are needed.
OBJECTIVE
In the absence of a data combining process, the cost of directly applying high-dimensional raw data to a deep classifier would be computationally expensive with regard to the response time, energy consumption, and memory requirement. Taking this into account, we aimed to develop a data fusion technique in a computationally efficient way to achieve a more comprehensive insight of human activity dynamics in a lower dimension. The major objective was considering statistical dependency of multisensory data and exploring intermodality correlation patterns for different activities.
METHODS
In this technique, the information in time (regardless of the number of sources) is transformed into a 2D space that facilitates classification of eating episodes from others. This is based on a hypothesis that data captured by various sensors are statistically associated with each other and the covariance matrix of all these signals has a unique distribution correlated with each activity which can be encoded on a contour representation. These representations are then used as input of a deep model to learn specific patterns associated with specific activity.
RESULTS
In order to show the generalizability of the proposed fusion algorithm, 2 different scenarios were taken into account. These scenarios were different in terms of temporal segment size, type of activity, wearable device, subjects, and deep learning architecture. The first scenario used a data set in which a single participant performed a limited number of activities while wearing the Empatica E4 wristband. In the second scenario, a data set related to the activities of daily living was used where 10 different participants wore inertial measurement units while performing a more complex set of activities. The precision metric obtained from leave-one-subject-out cross-validation for the second scenario reached 0.803. The impact of missing data on performance degradation was also evaluated.
CONCLUSIONS
To conclude, the proposed fusion technique provides the possibility of embedding joint variability information over different modalities in just a single 2D representation which results in obtaining a more global view of different aspects of daily human activities at hand, and yet preserving the desired performance level in activity recognition.
Identifiants
pubmed: 33507156
pii: v9i1e21926
doi: 10.2196/21926
pmc: PMC7878112
doi:
Types de publication
Journal Article
Research Support, Non-U.S. Gov't
Langues
eng
Sous-ensembles de citation
IM
Pagination
e21926Informations de copyright
©Nooshin Bahador, Denzil Ferreira, Satu Tamminen, Jukka Kortelainen. Originally published in JMIR mHealth and uHealth (http://mhealth.jmir.org), 28.01.2021.
Références
IEEE J Biomed Health Inform. 2015 Jan;19(1):282-9
pubmed: 24771599
IEEE Trans Inf Technol Biomed. 2011 Jan;15(1):74-82
pubmed: 21075728
Sci Rep. 2017 Jan 30;7:41690
pubmed: 28134303
IEEE J Biomed Health Inform. 2017 May;21(3):607-618
pubmed: 27834659
IEEE J Biomed Health Inform. 2021 Jan;25(1):22-34
pubmed: 32750897
NPJ Digit Med. 2020 Mar 13;3:38
pubmed: 32195373
Proc ACM Int Conf Ubiquitous Comput. 2015 Sep;2015:1029-1040
pubmed: 29520397
Data Brief. 2020 Aug 02;32:106122
pubmed: 32904359
IEEE J Biomed Health Inform. 2017 Nov;21(6):1495-1503
pubmed: 28113335
IEEE J Biomed Health Inform. 2014 Jan;18(1):278-89
pubmed: 24403426
Nutrients. 2019 May 24;11(5):
pubmed: 31137677
Annu Int Conf IEEE Eng Med Biol Soc. 2010;2010:1004-7
pubmed: 21096991
JMIR Mhealth Uhealth. 2019 Feb 07;7(2):e11201
pubmed: 30730297
Physiol Meas. 2014 May;35(5):739-51
pubmed: 24671094
Proc ACM Interact Mob Wearable Ubiquitous Technol. 2017 Sep;1(3):
pubmed: 30135957
JMIR Mhealth Uhealth. 2018 Sep 04;6(9):e170
pubmed: 30181111
IEEE J Biomed Health Inform. 2019 May;23(3):1022-1031
pubmed: 30040664
IEEE Sens J. 2012;12(5):1340-1348
pubmed: 22675270
IEEE J Biomed Health Inform. 2018 Jan;22(1):23-32
pubmed: 28463209
IEEE Trans Biomed Eng. 2012 Mar;59(3):687-96
pubmed: 22156943
IEEE Trans Biomed Eng. 2017 Sep;64(9):2075-2089
pubmed: 28092510
IEEE J Biomed Health Inform. 2018 May;22(3):678-685
pubmed: 28534801
IEEE Trans Biomed Eng. 2017 Jun;64(6):1369-1379
pubmed: 28113223
IEEE J Biomed Health Inform. 2020 Jun;24(6):1727-1737
pubmed: 31567103