Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks

Recently, convolutional neural networks (CNN) have been intensively investigated for the classification of remote sensing data by extracting invariant and abstract features suitable for classification. In this paper, a novel framework is proposed for the fusion of hyperspectral images and LiDAR-deri...

Full description

Bibliographic Details
Main Authors: Hao Li, Pedram Ghamisi, Uwe Soergel, Xiao Xiang Zhu
Format: Article
Language:English
Published: MDPI AG 2018-10-01
Series:Remote Sensing
Subjects:
Online Access:http://www.mdpi.com/2072-4292/10/10/1649
id doaj-765edb1c9cb14f3db7bd7d69452f84a1
record_format Article
spelling doaj-765edb1c9cb14f3db7bd7d69452f84a12020-11-24T21:43:14ZengMDPI AGRemote Sensing2072-42922018-10-011010164910.3390/rs10101649rs10101649Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural NetworksHao Li0Pedram Ghamisi1Uwe Soergel2Xiao Xiang Zhu3Signal Processing in Earth Observation (SiPEO), Technical University of Munich (TUM), Arcisstr. 21, 80333 Munich, GermanyHelmholtz-Zentrum Dresden-Rossendorf, Helmholtz Institute Freiberg for Resource Technology, Exploration, Chemnitzer Str. 40, D-09599 Freiberg, GermanyInstitute for Photogrammetry (ifp), University of Stuttgart, 70174 Stuttgart, GermanySignal Processing in Earth Observation (SiPEO), Technical University of Munich (TUM), Arcisstr. 21, 80333 Munich, GermanyRecently, convolutional neural networks (CNN) have been intensively investigated for the classification of remote sensing data by extracting invariant and abstract features suitable for classification. In this paper, a novel framework is proposed for the fusion of hyperspectral images and LiDAR-derived elevation data based on CNN and composite kernels. First, extinction profiles are applied to both data sources in order to extract spatial and elevation features from hyperspectral and LiDAR-derived data, respectively. Second, a three-stream CNN is designed to extract informative spectral, spatial, and elevation features individually from both available sources. The combination of extinction profiles and CNN features enables us to jointly benefit from low-level and high-level features to improve classification performance. To fuse the heterogeneous spectral, spatial, and elevation features extracted by CNN, instead of a simple stacking strategy, a multi-sensor composite kernels (MCK) scheme is designed. This scheme helps us to achieve higher spectral, spatial, and elevation separability of the extracted features and effectively perform multi-sensor data fusion in kernel space. In this context, a support vector machine and extreme learning machine with their composite kernels version are employed to produce the final classification result. The proposed framework is carried out on two widely used data sets with different characteristics: an urban data set captured over Houston, USA, and a rural data set captured over Trento, Italy. The proposed framework yields the highest OA of 92 . 57 % and 97 . 91 % for Houston and Trento data sets. Experimental results confirm that the proposed fusion framework can produce competitive results in both urban and rural areas in terms of classification accuracy, and significantly mitigate the salt and pepper noise in classification maps.http://www.mdpi.com/2072-4292/10/10/1649data fusionextinction profiles (EPs)composite kernelsconvolutional neural networks (CNN)feature extraction (FE)
collection DOAJ
language English
format Article
sources DOAJ
author Hao Li
Pedram Ghamisi
Uwe Soergel
Xiao Xiang Zhu
spellingShingle Hao Li
Pedram Ghamisi
Uwe Soergel
Xiao Xiang Zhu
Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
Remote Sensing
data fusion
extinction profiles (EPs)
composite kernels
convolutional neural networks (CNN)
feature extraction (FE)
author_facet Hao Li
Pedram Ghamisi
Uwe Soergel
Xiao Xiang Zhu
author_sort Hao Li
title Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
title_short Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
title_full Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
title_fullStr Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
title_full_unstemmed Hyperspectral and LiDAR Fusion Using Deep Three-Stream Convolutional Neural Networks
title_sort hyperspectral and lidar fusion using deep three-stream convolutional neural networks
publisher MDPI AG
series Remote Sensing
issn 2072-4292
publishDate 2018-10-01
description Recently, convolutional neural networks (CNN) have been intensively investigated for the classification of remote sensing data by extracting invariant and abstract features suitable for classification. In this paper, a novel framework is proposed for the fusion of hyperspectral images and LiDAR-derived elevation data based on CNN and composite kernels. First, extinction profiles are applied to both data sources in order to extract spatial and elevation features from hyperspectral and LiDAR-derived data, respectively. Second, a three-stream CNN is designed to extract informative spectral, spatial, and elevation features individually from both available sources. The combination of extinction profiles and CNN features enables us to jointly benefit from low-level and high-level features to improve classification performance. To fuse the heterogeneous spectral, spatial, and elevation features extracted by CNN, instead of a simple stacking strategy, a multi-sensor composite kernels (MCK) scheme is designed. This scheme helps us to achieve higher spectral, spatial, and elevation separability of the extracted features and effectively perform multi-sensor data fusion in kernel space. In this context, a support vector machine and extreme learning machine with their composite kernels version are employed to produce the final classification result. The proposed framework is carried out on two widely used data sets with different characteristics: an urban data set captured over Houston, USA, and a rural data set captured over Trento, Italy. The proposed framework yields the highest OA of 92 . 57 % and 97 . 91 % for Houston and Trento data sets. Experimental results confirm that the proposed fusion framework can produce competitive results in both urban and rural areas in terms of classification accuracy, and significantly mitigate the salt and pepper noise in classification maps.
topic data fusion
extinction profiles (EPs)
composite kernels
convolutional neural networks (CNN)
feature extraction (FE)
url http://www.mdpi.com/2072-4292/10/10/1649
work_keys_str_mv AT haoli hyperspectralandlidarfusionusingdeepthreestreamconvolutionalneuralnetworks
AT pedramghamisi hyperspectralandlidarfusionusingdeepthreestreamconvolutionalneuralnetworks
AT uwesoergel hyperspectralandlidarfusionusingdeepthreestreamconvolutionalneuralnetworks
AT xiaoxiangzhu hyperspectralandlidarfusionusingdeepthreestreamconvolutionalneuralnetworks
_version_ 1725914756630446080