Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception

Hearing-impaired people often struggle to follow the speech stream of an individual talker in noisy environments. Recent studies show that the brain tracks attended speech and that the attended talker can be decoded from neural data on a single-trial level. This raises the possibility of “neuro-stee...

Full description

Bibliographic Details
Main Authors: Enea Ceolini, Jens Hjortkjær, Daniel D.E. Wong, James O’Sullivan, Vinay S. Raghavan, Jose Herrero, Ashesh D. Mehta, Shih-Chii Liu, Nima Mesgarani
Format: Article
Language:English
Published: Elsevier 2020-12-01
Series:NeuroImage
Subjects:
EEG
Online Access:http://www.sciencedirect.com/science/article/pii/S1053811920307680
id doaj-6d22285e1b364afd8312f6db12b1c33f
record_format Article
spelling doaj-6d22285e1b364afd8312f6db12b1c33f2020-11-25T03:40:11ZengElsevierNeuroImage1095-95722020-12-01223117282Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perceptionEnea Ceolini0Jens Hjortkjær1Daniel D.E. Wong2James O’Sullivan3Vinay S. Raghavan4Jose Herrero5Ashesh D. Mehta6Shih-Chii Liu7Nima Mesgarani8Corresponding authors.; University of Zürich and ETH Zürich, Institute of Neuroinformatics, SwitzerlandDepartment of Health Technology, Danmarks Tekniske Universitet DTU, Kongens Lyngby, Denmark; Danish Research Centre for Magnetic Resonance, Copenhagen University Hospital Hvidovre, Hvidovre, DenmarkLaboratoire des Systèmes Perceptifs, CNRS, UMR 8248, Paris, France; Département d’Études Cognitives, École Normale Supérieure, PSL Research University, Paris, FranceDepartment of Electrical Engineering, Columbia University, New York, NY, USA; Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USADepartment of Electrical Engineering, Columbia University, New York, NY, USA; Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USADepartment of Neurosurgery, Hofstra-Northwell School of Medicine and Feinstein Institute for Medical Research, Manhasset, New York, NY, USADepartment of Neurosurgery, Hofstra-Northwell School of Medicine and Feinstein Institute for Medical Research, Manhasset, New York, NY, USAUniversity of Zürich and ETH Zürich, Institute of Neuroinformatics, SwitzerlandCorresponding authors.; Department of Electrical Engineering, Columbia University, New York, NY, USA; Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USAHearing-impaired people often struggle to follow the speech stream of an individual talker in noisy environments. Recent studies show that the brain tracks attended speech and that the attended talker can be decoded from neural data on a single-trial level. This raises the possibility of “neuro-steered” hearing devices in which the brain-decoded intention of a hearing-impaired listener is used to enhance the voice of the attended speaker from a speech separation front-end. So far, methods that use this paradigm have focused on optimizing the brain decoding and the acoustic speech separation independently. In this work, we propose a novel framework called brain-informed speech separation (BISS)11 BISS: brain-informed speech separation. in which the information about the attended speech, as decoded from the subject’s brain, is directly used to perform speech separation in the front-end. We present a deep learning model that uses neural data to extract the clean audio signal that a listener is attending to from a multi-talker speech mixture. We show that the framework can be applied successfully to the decoded output from either invasive intracranial electroencephalography (iEEG) or non-invasive electroencephalography (EEG) recordings from hearing-impaired subjects. It also results in improved speech separation, even in scenes with background noise. The generalization capability of the system renders it a perfect candidate for neuro-steered hearing-assistive devices.http://www.sciencedirect.com/science/article/pii/S1053811920307680EEGNeuro-steeredCognitive controlSpeech separationDeep learningHearing aid
collection DOAJ
language English
format Article
sources DOAJ
author Enea Ceolini
Jens Hjortkjær
Daniel D.E. Wong
James O’Sullivan
Vinay S. Raghavan
Jose Herrero
Ashesh D. Mehta
Shih-Chii Liu
Nima Mesgarani
spellingShingle Enea Ceolini
Jens Hjortkjær
Daniel D.E. Wong
James O’Sullivan
Vinay S. Raghavan
Jose Herrero
Ashesh D. Mehta
Shih-Chii Liu
Nima Mesgarani
Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
NeuroImage
EEG
Neuro-steered
Cognitive control
Speech separation
Deep learning
Hearing aid
author_facet Enea Ceolini
Jens Hjortkjær
Daniel D.E. Wong
James O’Sullivan
Vinay S. Raghavan
Jose Herrero
Ashesh D. Mehta
Shih-Chii Liu
Nima Mesgarani
author_sort Enea Ceolini
title Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
title_short Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
title_full Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
title_fullStr Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
title_full_unstemmed Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception
title_sort brain-informed speech separation (biss) for enhancement of target speaker in multitalker speech perception
publisher Elsevier
series NeuroImage
issn 1095-9572
publishDate 2020-12-01
description Hearing-impaired people often struggle to follow the speech stream of an individual talker in noisy environments. Recent studies show that the brain tracks attended speech and that the attended talker can be decoded from neural data on a single-trial level. This raises the possibility of “neuro-steered” hearing devices in which the brain-decoded intention of a hearing-impaired listener is used to enhance the voice of the attended speaker from a speech separation front-end. So far, methods that use this paradigm have focused on optimizing the brain decoding and the acoustic speech separation independently. In this work, we propose a novel framework called brain-informed speech separation (BISS)11 BISS: brain-informed speech separation. in which the information about the attended speech, as decoded from the subject’s brain, is directly used to perform speech separation in the front-end. We present a deep learning model that uses neural data to extract the clean audio signal that a listener is attending to from a multi-talker speech mixture. We show that the framework can be applied successfully to the decoded output from either invasive intracranial electroencephalography (iEEG) or non-invasive electroencephalography (EEG) recordings from hearing-impaired subjects. It also results in improved speech separation, even in scenes with background noise. The generalization capability of the system renders it a perfect candidate for neuro-steered hearing-assistive devices.
topic EEG
Neuro-steered
Cognitive control
Speech separation
Deep learning
Hearing aid
url http://www.sciencedirect.com/science/article/pii/S1053811920307680
work_keys_str_mv AT eneaceolini braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT jenshjortkjær braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT danieldewong braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT jamesosullivan braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT vinaysraghavan braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT joseherrero braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT asheshdmehta braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT shihchiiliu braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
AT nimamesgarani braininformedspeechseparationbissforenhancementoftargetspeakerinmultitalkerspeechperception
_version_ 1724535728802627584