Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors

We propose an unmixing framework for enhancing endmember fraction maps using a combination of spectral and visible images. The new method, data fusion through spatial information-aided learning (DFuSIAL), is based on a learning process for the fusion of a multispectral image of low spatial resolutio...

Full description

Bibliographic Details
Main Authors: Fadi Kizel, Jón Atli Benediktsson
Format: Article
Language:English
Published: MDPI AG 2020-04-01
Series:Remote Sensing
Subjects:
Online Access:https://www.mdpi.com/2072-4292/12/8/1255
id doaj-4df6c901d3774353a2eee7fbabb6a494
record_format Article
spelling doaj-4df6c901d3774353a2eee7fbabb6a4942020-11-25T02:02:55ZengMDPI AGRemote Sensing2072-42922020-04-01121255125510.3390/rs12081255Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different SensorsFadi Kizel0Jón Atli Benediktsson1Department of Mapping and Geoinformation Engineering, Civil and Environmental Engineering, Technion-Israel Institute of Technology,32000 Haifa, IsraelFaculty of Electrical and Computer Engineering, University of Iceland, 102 Reykjavík, IcelandWe propose an unmixing framework for enhancing endmember fraction maps using a combination of spectral and visible images. The new method, data fusion through spatial information-aided learning (DFuSIAL), is based on a learning process for the fusion of a multispectral image of low spatial resolution and a visible RGB image of high spatial resolution. Unlike commonly used methods, DFuSIAL allows for fusing data from different sensors. To achieve this objective, we apply a learning process using automatically extracted invariant points, which are assumed to have the same land cover type in both images. First, we estimate the fraction maps of a set of endmembers for the spectral image. Then, we train a spatial-features aided neural network (SFFAN) to learn the relationship between the fractions, the visible bands, and rotation-invariant spatial features for learning (RISFLs) that we extract from the RGB image. Our experiments show that the proposed DFuSIAL method obtains fraction maps with significantly enhanced spatial resolution and an average mean absolute error between 2% and 4% compared to the reference ground truth. Furthermore, it is shown that the proposed method is preferable to other examined state-of-the-art methods, especially when data is obtained from different instruments and in cases with missing-data pixels.https://www.mdpi.com/2072-4292/12/8/1255remote sensingspectral unmixingmultispectral imagesdata fusionspatial resolutionspatial information
collection DOAJ
language English
format Article
sources DOAJ
author Fadi Kizel
Jón Atli Benediktsson
spellingShingle Fadi Kizel
Jón Atli Benediktsson
Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
Remote Sensing
remote sensing
spectral unmixing
multispectral images
data fusion
spatial resolution
spatial information
author_facet Fadi Kizel
Jón Atli Benediktsson
author_sort Fadi Kizel
title Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
title_short Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
title_full Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
title_fullStr Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
title_full_unstemmed Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
title_sort spatially enhanced spectral unmixing through data fusion of spectral and visible images from different sensors
publisher MDPI AG
series Remote Sensing
issn 2072-4292
publishDate 2020-04-01
description We propose an unmixing framework for enhancing endmember fraction maps using a combination of spectral and visible images. The new method, data fusion through spatial information-aided learning (DFuSIAL), is based on a learning process for the fusion of a multispectral image of low spatial resolution and a visible RGB image of high spatial resolution. Unlike commonly used methods, DFuSIAL allows for fusing data from different sensors. To achieve this objective, we apply a learning process using automatically extracted invariant points, which are assumed to have the same land cover type in both images. First, we estimate the fraction maps of a set of endmembers for the spectral image. Then, we train a spatial-features aided neural network (SFFAN) to learn the relationship between the fractions, the visible bands, and rotation-invariant spatial features for learning (RISFLs) that we extract from the RGB image. Our experiments show that the proposed DFuSIAL method obtains fraction maps with significantly enhanced spatial resolution and an average mean absolute error between 2% and 4% compared to the reference ground truth. Furthermore, it is shown that the proposed method is preferable to other examined state-of-the-art methods, especially when data is obtained from different instruments and in cases with missing-data pixels.
topic remote sensing
spectral unmixing
multispectral images
data fusion
spatial resolution
spatial information
url https://www.mdpi.com/2072-4292/12/8/1255
work_keys_str_mv AT fadikizel spatiallyenhancedspectralunmixingthroughdatafusionofspectralandvisibleimagesfromdifferentsensors
AT jonatlibenediktsson spatiallyenhancedspectralunmixingthroughdatafusionofspectralandvisibleimagesfromdifferentsensors
_version_ 1724950648378621952