Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing

Human-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data c...

Full description

Bibliographic Details
Main Authors: Jiancun Zhou, Tao Xu, Sheng Ren, Kehua Guo
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9094630/
id doaj-1800cd82f4024d0a832f5b086d2eb42b
record_format Article
spelling doaj-1800cd82f4024d0a832f5b086d2eb42b2021-03-30T02:15:20ZengIEEEIEEE Access2169-35362020-01-018967279673710.1109/ACCESS.2020.29952689094630Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd SensingJiancun Zhou0Tao Xu1Sheng Ren2Kehua Guo3https://orcid.org/0000-0003-4143-6399All-Solid-State Energy Storage Materials and Devices Key Laboratory of Hunan Province, Hunan City University, Yiyang, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaSchool of Computer Science and Engineering, Central South University, Changsha, ChinaHuman-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data contain unique information and attributes, which can be complementary. Combining data from many different modalities will get more information. However, current deep learning is usually only for bimodal data. In order for artificial intelligence to make further breakthroughs in understanding our real world, it needs to be able to process data in different modalities together. The key step is to be able to map these different modalities data into the same space. In order to process multimodal data better, we propose a fusion and classification method for multimodal data. First, a multimodal data space is constructed, and data of different modalities are mapped into the multimodal data space to obtain a unified representation of different modalities data. Then, through bilinear pooling, the representations of different modality are fused, and the fused vectors are used in the classification task. Through the experimental verification on the multi-modal data set, it proves that the multi-modal fusion representation is effective, and the classification effect is more accurate than the single-modal data.https://ieeexplore.ieee.org/document/9094630/Multimodal dataunified representationmobile crowd sensinghuman-driven edge computing
collection DOAJ
language English
format Article
sources DOAJ
author Jiancun Zhou
Tao Xu
Sheng Ren
Kehua Guo
spellingShingle Jiancun Zhou
Tao Xu
Sheng Ren
Kehua Guo
Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
IEEE Access
Multimodal data
unified representation
mobile crowd sensing
human-driven edge computing
author_facet Jiancun Zhou
Tao Xu
Sheng Ren
Kehua Guo
author_sort Jiancun Zhou
title Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
title_short Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
title_full Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
title_fullStr Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
title_full_unstemmed Two-Stage Spatial Mapping for Multimodal Data Fusion in Mobile Crowd Sensing
title_sort two-stage spatial mapping for multimodal data fusion in mobile crowd sensing
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Human-driven Edge Computing (HEC) integrates the elements of humans, devices, Internet and information, and mobile crowd sensing become an important means of data collection. In HEC, the data collected from large-scale sensing usually includes a variety of modalities. These different modality data contain unique information and attributes, which can be complementary. Combining data from many different modalities will get more information. However, current deep learning is usually only for bimodal data. In order for artificial intelligence to make further breakthroughs in understanding our real world, it needs to be able to process data in different modalities together. The key step is to be able to map these different modalities data into the same space. In order to process multimodal data better, we propose a fusion and classification method for multimodal data. First, a multimodal data space is constructed, and data of different modalities are mapped into the multimodal data space to obtain a unified representation of different modalities data. Then, through bilinear pooling, the representations of different modality are fused, and the fused vectors are used in the classification task. Through the experimental verification on the multi-modal data set, it proves that the multi-modal fusion representation is effective, and the classification effect is more accurate than the single-modal data.
topic Multimodal data
unified representation
mobile crowd sensing
human-driven edge computing
url https://ieeexplore.ieee.org/document/9094630/
work_keys_str_mv AT jiancunzhou twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing
AT taoxu twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing
AT shengren twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing
AT kehuaguo twostagespatialmappingformultimodaldatafusioninmobilecrowdsensing
_version_ 1724185493685403648