Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.

Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number o...

Full description

Bibliographic Details
Main Authors: Florent Bocquelet, Thomas Hueber, Laurent Girin, Christophe Savariaux, Blaise Yvert
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2016-11-01
Series:PLoS Computational Biology
Online Access:http://europepmc.org/articles/PMC5120792?pdf=render
id doaj-032eacf1c53a44c488b9e1a8dbca69f0
record_format Article
spelling doaj-032eacf1c53a44c488b9e1a8dbca69f02020-11-24T21:58:58ZengPublic Library of Science (PLoS)PLoS Computational Biology1553-734X1553-73582016-11-011211e100511910.1371/journal.pcbi.1005119Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.Florent BocqueletThomas HueberLaurent GirinChristophe SavariauxBlaise YvertRestoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer.http://europepmc.org/articles/PMC5120792?pdf=render
collection DOAJ
language English
format Article
sources DOAJ
author Florent Bocquelet
Thomas Hueber
Laurent Girin
Christophe Savariaux
Blaise Yvert
spellingShingle Florent Bocquelet
Thomas Hueber
Laurent Girin
Christophe Savariaux
Blaise Yvert
Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
PLoS Computational Biology
author_facet Florent Bocquelet
Thomas Hueber
Laurent Girin
Christophe Savariaux
Blaise Yvert
author_sort Florent Bocquelet
title Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
title_short Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
title_full Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
title_fullStr Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
title_full_unstemmed Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
title_sort real-time control of an articulatory-based speech synthesizer for brain computer interfaces.
publisher Public Library of Science (PLoS)
series PLoS Computational Biology
issn 1553-734X
1553-7358
publishDate 2016-11-01
description Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer.
url http://europepmc.org/articles/PMC5120792?pdf=render
work_keys_str_mv AT florentbocquelet realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces
AT thomashueber realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces
AT laurentgirin realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces
AT christophesavariaux realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces
AT blaiseyvert realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces
_version_ 1725849992144355328