EEG Classification of Motor Imagery Using a Novel Deep Learning Framework

Successful applications of brain-computer interface (BCI) approaches to motor imagery (MI) are still limited. In this paper, we propose a classification framework for MI electroencephalogram (EEG) signals that combines a convolutional neural network (CNN) architecture with a variational autoencoder...

Full description

Bibliographic Details
Main Authors: Mengxi Dai, Dezhi Zheng, Rui Na, Shuai Wang, Shuailei Zhang
Format: Article
Language:English
Published: MDPI AG 2019-01-01
Series:Sensors
Subjects:
EEG
Online Access:https://www.mdpi.com/1424-8220/19/3/551
id doaj-3deb711b3518446eb389c0dc89419f63
record_format Article
spelling doaj-3deb711b3518446eb389c0dc89419f632020-11-24T21:48:34ZengMDPI AGSensors1424-82202019-01-0119355110.3390/s19030551s19030551EEG Classification of Motor Imagery Using a Novel Deep Learning FrameworkMengxi Dai0Dezhi Zheng1Rui Na2Shuai Wang3Shuailei Zhang4School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, ChinaSchool of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, ChinaSchool of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, ChinaSchool of Computer Science and Engineering, Beihang University, Beijing 100191, ChinaSchool of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, ChinaSuccessful applications of brain-computer interface (BCI) approaches to motor imagery (MI) are still limited. In this paper, we propose a classification framework for MI electroencephalogram (EEG) signals that combines a convolutional neural network (CNN) architecture with a variational autoencoder (VAE) for classification. The decoder of the VAE generates a Gaussian distribution, so it can be used to fit the Gaussian distribution of EEG signals. A new representation of input was developed by combining the time, frequency, and channel information from the EEG signal, and the CNN-VAE method was designed and optimized accordingly for this form of input. In this network, the classification of the extracted CNN features is performed via the deep network VAE. Our framework, with an average kappa value of 0.564, outperforms the best classification method in the literature for BCI Competition IV dataset 2b with a 3% improvement. Furthermore, using our own dataset, the CNN-VAE framework also yields the best performance for both three-electrode and five-electrode EEGs and achieves the best average kappa values 0.568 and 0.603, respectively. Our results show that the proposed CNN-VAE method raises performance to the current state of the art.https://www.mdpi.com/1424-8220/19/3/551EEGdeep learningshort-time Fourier transformconvolutional neural networkvariational autoencoder
collection DOAJ
language English
format Article
sources DOAJ
author Mengxi Dai
Dezhi Zheng
Rui Na
Shuai Wang
Shuailei Zhang
spellingShingle Mengxi Dai
Dezhi Zheng
Rui Na
Shuai Wang
Shuailei Zhang
EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
Sensors
EEG
deep learning
short-time Fourier transform
convolutional neural network
variational autoencoder
author_facet Mengxi Dai
Dezhi Zheng
Rui Na
Shuai Wang
Shuailei Zhang
author_sort Mengxi Dai
title EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
title_short EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
title_full EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
title_fullStr EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
title_full_unstemmed EEG Classification of Motor Imagery Using a Novel Deep Learning Framework
title_sort eeg classification of motor imagery using a novel deep learning framework
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2019-01-01
description Successful applications of brain-computer interface (BCI) approaches to motor imagery (MI) are still limited. In this paper, we propose a classification framework for MI electroencephalogram (EEG) signals that combines a convolutional neural network (CNN) architecture with a variational autoencoder (VAE) for classification. The decoder of the VAE generates a Gaussian distribution, so it can be used to fit the Gaussian distribution of EEG signals. A new representation of input was developed by combining the time, frequency, and channel information from the EEG signal, and the CNN-VAE method was designed and optimized accordingly for this form of input. In this network, the classification of the extracted CNN features is performed via the deep network VAE. Our framework, with an average kappa value of 0.564, outperforms the best classification method in the literature for BCI Competition IV dataset 2b with a 3% improvement. Furthermore, using our own dataset, the CNN-VAE framework also yields the best performance for both three-electrode and five-electrode EEGs and achieves the best average kappa values 0.568 and 0.603, respectively. Our results show that the proposed CNN-VAE method raises performance to the current state of the art.
topic EEG
deep learning
short-time Fourier transform
convolutional neural network
variational autoencoder
url https://www.mdpi.com/1424-8220/19/3/551
work_keys_str_mv AT mengxidai eegclassificationofmotorimageryusinganoveldeeplearningframework
AT dezhizheng eegclassificationofmotorimageryusinganoveldeeplearningframework
AT ruina eegclassificationofmotorimageryusinganoveldeeplearningframework
AT shuaiwang eegclassificationofmotorimageryusinganoveldeeplearningframework
AT shuaileizhang eegclassificationofmotorimageryusinganoveldeeplearningframework
_version_ 1725891458715615232