Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
Human-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data c...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9094630/ |
id |
doaj-1800cd82f4024d0a832f5b086d2eb42b |
---|---|
record_format |
Article |
spelling |
doaj-1800cd82f4024d0a832f5b086d2eb42b2021-03-30T02:15:20ZengIEEEIEEE Access2169-35362020-01-018967279673710.1109/ACCESS.2020.29952689094630Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd SensingJiancun Zhou0Tao Xu1Sheng Ren2Kehua Guo3https://orcid.org/0000-0003-4143-6399All-Solid-State Energy Storage Materials and Devices Key Laboratory of Hunan Province, Hunan City University, Yiyang, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaHuman-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data contain unique information and attributes, which can be complementary. Combining data from many different modalities will get more information. However, current deep learning is usually only for bimodal data. In order for artificial intelligence to make further breakthroughs in understanding our real world, it needs to be able to process data in different modalities together. The key step is to be able to map these different modalities data into the same space. In order to process multimodal data better, we propose a fusion and classification method for multimodal data. First, a multimodal data space is constructed, and data of different modalities are mapped into the multimodal data space to obtain a unified representation of different modalities data. Then, through bilinear pooling, the representations of different modality are fused, and the fused vectors are used in the classification task. Through the experimental verification on the multi-modal data set, it proves that the multi-modal fusion representation is effective, and the classification effect is more accurate than the single-modal data.https://ieeexplore.ieee.org/document/9094630/Multimodal dataunified representationmobile crowd sensinghuman-driven edge computing |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Jiancun Zhou Tao Xu Sheng Ren Kehua Guo |
spellingShingle |
Jiancun Zhou Tao Xu Sheng Ren Kehua Guo Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing IEEE Access Multimodal data unified representation mobile crowd sensing human-driven edge computing |
author_facet |
Jiancun Zhou Tao Xu Sheng Ren Kehua Guo |
author_sort |
Jiancun Zhou |
title |
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing |
title_short |
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing |
title_full |
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing |
title_fullStr |
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing |
title_full_unstemmed |
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing |
title_sort |
two-stage spatial mapping for multimodal data fusion in mobile crowd sensing |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
Human-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data contain unique information and attributes, which can be complementary. Combining data from many different modalities will get more information. However, current deep learning is usually only for bimodal data. In order for artificial intelligence to make further breakthroughs in understanding our real world, it needs to be able to process data in different modalities together. The key step is to be able to map these different modalities data into the same space. In order to process multimodal data better, we propose a fusion and classification method for multimodal data. First, a multimodal data space is constructed, and data of different modalities are mapped into the multimodal data space to obtain a unified representation of different modalities data. Then, through bilinear pooling, the representations of different modality are fused, and the fused vectors are used in the classification task. Through the experimental verification on the multi-modal data set, it proves that the multi-modal fusion representation is effective, and the classification effect is more accurate than the single-modal data. |
topic |
Multimodal data unified representation mobile crowd sensing human-driven edge computing |
url |
https://ieeexplore.ieee.org/document/9094630/ |
work_keys_str_mv |
AT jiancunzhou twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing AT taoxu twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing AT shengren twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing AT kehuaguo twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing |
_version_ |
1724185493685403648 |