Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification

Brain–computer interfaces (BCIs) can help people with limited motor abilities to interact with their environment without external assistance. A major challenge in electroencephalogram (EEG)-based BCI development and research is the cross-subject classification of motor imagery data. Due to the highl...

Full description

Bibliographic Details
Main Authors: Karel Roots, Yar Muhammad, Naveed Muhammad
Format: Article
Language:English
Published: MDPI AG 2020-09-01
Series:Computers
Subjects:
Online Access:https://www.mdpi.com/2073-431X/9/3/72
id doaj-dd29a92fb9b244c39d503cbc799d57dd
record_format Article
spelling doaj-dd29a92fb9b244c39d503cbc799d57dd2020-11-25T02:30:50ZengMDPI AGComputers2073-431X2020-09-019727210.3390/computers9030072Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery ClassificationKarel Roots0Yar Muhammad1Naveed Muhammad2Institute of Computer Science, University of Tartu, 51009 Tartu, EstoniaInstitute of Computer Science, University of Tartu, 51009 Tartu, EstoniaInstitute of Computer Science, University of Tartu, 51009 Tartu, EstoniaBrain–computer interfaces (BCIs) can help people with limited motor abilities to interact with their environment without external assistance. A major challenge in electroencephalogram (EEG)-based BCI development and research is the cross-subject classification of motor imagery data. Due to the highly individualized nature of EEG signals, it has been difficult to develop a cross-subject classification method that achieves sufficiently high accuracy when predicting the subject’s intention. In this study, we propose a multi-branch 2D convolutional neural network (CNN) that utilizes different hyperparameter values for each branch and is more flexible to data from different subjects. Our model, EEGNet Fusion, achieves 84.1% and 83.8% accuracy when tested on the 103-subject <i>eegmmidb</i> dataset for executed and imagined motor actions, respectively. The model achieved statistically significantly higher results compared with three state-of-the-art CNN classifiers: EEGNet, ShallowConvNet, and DeepConvNet. However, the computational cost of the proposed model is up to four times higher than the model with the lowest computational cost used for comparison.https://www.mdpi.com/2073-431X/9/3/72brain–computer interface (BCI)convolutional neural network (CNN)deep learningelectroencephalography (EEG)fusion networkmotor imagery (MI)
collection DOAJ
language English
format Article
sources DOAJ
author Karel Roots
Yar Muhammad
Naveed Muhammad
spellingShingle Karel Roots
Yar Muhammad
Naveed Muhammad
Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
Computers
brain–computer interface (BCI)
convolutional neural network (CNN)
deep learning
electroencephalography (EEG)
fusion network
motor imagery (MI)
author_facet Karel Roots
Yar Muhammad
Naveed Muhammad
author_sort Karel Roots
title Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
title_short Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
title_full Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
title_fullStr Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
title_full_unstemmed Fusion Convolutional Neural Network for Cross-Subject EEG Motor Imagery Classification
title_sort fusion convolutional neural network for cross-subject eeg motor imagery classification
publisher MDPI AG
series Computers
issn 2073-431X
publishDate 2020-09-01
description Brain–computer interfaces (BCIs) can help people with limited motor abilities to interact with their environment without external assistance. A major challenge in electroencephalogram (EEG)-based BCI development and research is the cross-subject classification of motor imagery data. Due to the highly individualized nature of EEG signals, it has been difficult to develop a cross-subject classification method that achieves sufficiently high accuracy when predicting the subject’s intention. In this study, we propose a multi-branch 2D convolutional neural network (CNN) that utilizes different hyperparameter values for each branch and is more flexible to data from different subjects. Our model, EEGNet Fusion, achieves 84.1% and 83.8% accuracy when tested on the 103-subject <i>eegmmidb</i> dataset for executed and imagined motor actions, respectively. The model achieved statistically significantly higher results compared with three state-of-the-art CNN classifiers: EEGNet, ShallowConvNet, and DeepConvNet. However, the computational cost of the proposed model is up to four times higher than the model with the lowest computational cost used for comparison.
topic brain–computer interface (BCI)
convolutional neural network (CNN)
deep learning
electroencephalography (EEG)
fusion network
motor imagery (MI)
url https://www.mdpi.com/2073-431X/9/3/72
work_keys_str_mv AT karelroots fusionconvolutionalneuralnetworkforcrosssubjecteegmotorimageryclassification
AT yarmuhammad fusionconvolutionalneuralnetworkforcrosssubjecteegmotorimageryclassification
AT naveedmuhammad fusionconvolutionalneuralnetworkforcrosssubjecteegmotorimageryclassification
_version_ 1724827475334135808