Decoding phonation with artificial intelligence (DeP AI): Proof of concept

Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this...

Full description

Bibliographic Details
Main Authors: Maria E. Powell, Marcelino Rodriguez Cancio, David Young, William Nock, Beshoy Abdelmessih, Amy Zeller, Irvin Perez Morales, Peng Zhang, C. Gaelyn Garrett, Douglas Schmidt, Jules White, Alexander Gelbard
Format: Article
Language:English
Published: Wiley 2019-06-01
Series:Laryngoscope Investigative Otolaryngology
Subjects:
Online Access:https://doi.org/10.1002/lio2.259
id doaj-9a0bbc3973284e13a5f811f602615336
record_format Article
spelling doaj-9a0bbc3973284e13a5f811f6026153362020-11-25T02:59:57ZengWileyLaryngoscope Investigative Otolaryngology2378-80382019-06-014332833410.1002/lio2.259Decoding phonation with artificial intelligence (DeP AI): Proof of conceptMaria E. Powell0Marcelino Rodriguez Cancio1David Young2William Nock3Beshoy Abdelmessih4Amy Zeller5Irvin Perez Morales6Peng Zhang7C. Gaelyn Garrett8Douglas Schmidt9Jules White10Alexander Gelbard11Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Information Technology Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Center of Research in Computational and Numerical Methods in Engineering Central University Marta Abreu of Las Villas Santa Clara CubaDepartment of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this study is to provide a proof‐of‐concept that embedded data within human phonation can be accurately and efficiently decoded with deep learning neural network analysis to differentiate between normal and disordered voices. Methods Acoustic recordings from 10 vocally‐healthy speakers, as well as 70 patients with one of seven voice disorders (n = 10 per diagnosis), were acquired from a clinical database. Acoustic signals were converted into spectrograms and used to train a convolutional neural network developed with the Keras library. The network architecture was trained separately for each of the seven diagnostic categories. Binary classification tasks (ie, to classify normal vs. disordered) were performed for each of the seven diagnostic categories. All models were validated using the 10‐fold cross‐validation technique. Results Binary classification averaged accuracies ranged from 58% to 90%. Models were most accurate in their classification of adductor spasmodic dysphonia, unilateral vocal fold paralysis, vocal fold polyp, polypoid corditis, and recurrent respiratory papillomatosis. Despite a small sample size, these findings are consistent with previously published data utilizing deep neural networks for classification of voice disorders. Conclusion Promising preliminary results support further study of deep neural networks for clinical detection and diagnosis of human voice disorders. Current models should be optimized with a larger sample size. Levels of Evidence Level IIIhttps://doi.org/10.1002/lio2.259Voice disordersdetectionacoustic analysisconvolutional neural networkclassification
collection DOAJ
language English
format Article
sources DOAJ
author Maria E. Powell
Marcelino Rodriguez Cancio
David Young
William Nock
Beshoy Abdelmessih
Amy Zeller
Irvin Perez Morales
Peng Zhang
C. Gaelyn Garrett
Douglas Schmidt
Jules White
Alexander Gelbard
spellingShingle Maria E. Powell
Marcelino Rodriguez Cancio
David Young
William Nock
Beshoy Abdelmessih
Amy Zeller
Irvin Perez Morales
Peng Zhang
C. Gaelyn Garrett
Douglas Schmidt
Jules White
Alexander Gelbard
Decoding phonation with artificial intelligence (DeP AI): Proof of concept
Laryngoscope Investigative Otolaryngology
Voice disorders
detection
acoustic analysis
convolutional neural network
classification
author_facet Maria E. Powell
Marcelino Rodriguez Cancio
David Young
William Nock
Beshoy Abdelmessih
Amy Zeller
Irvin Perez Morales
Peng Zhang
C. Gaelyn Garrett
Douglas Schmidt
Jules White
Alexander Gelbard
author_sort Maria E. Powell
title Decoding phonation with artificial intelligence (DeP AI): Proof of concept
title_short Decoding phonation with artificial intelligence (DeP AI): Proof of concept
title_full Decoding phonation with artificial intelligence (DeP AI): Proof of concept
title_fullStr Decoding phonation with artificial intelligence (DeP AI): Proof of concept
title_full_unstemmed Decoding phonation with artificial intelligence (DeP AI): Proof of concept
title_sort decoding phonation with artificial intelligence (dep ai): proof of concept
publisher Wiley
series Laryngoscope Investigative Otolaryngology
issn 2378-8038
publishDate 2019-06-01
description Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this study is to provide a proof‐of‐concept that embedded data within human phonation can be accurately and efficiently decoded with deep learning neural network analysis to differentiate between normal and disordered voices. Methods Acoustic recordings from 10 vocally‐healthy speakers, as well as 70 patients with one of seven voice disorders (n = 10 per diagnosis), were acquired from a clinical database. Acoustic signals were converted into spectrograms and used to train a convolutional neural network developed with the Keras library. The network architecture was trained separately for each of the seven diagnostic categories. Binary classification tasks (ie, to classify normal vs. disordered) were performed for each of the seven diagnostic categories. All models were validated using the 10‐fold cross‐validation technique. Results Binary classification averaged accuracies ranged from 58% to 90%. Models were most accurate in their classification of adductor spasmodic dysphonia, unilateral vocal fold paralysis, vocal fold polyp, polypoid corditis, and recurrent respiratory papillomatosis. Despite a small sample size, these findings are consistent with previously published data utilizing deep neural networks for classification of voice disorders. Conclusion Promising preliminary results support further study of deep neural networks for clinical detection and diagnosis of human voice disorders. Current models should be optimized with a larger sample size. Levels of Evidence Level III
topic Voice disorders
detection
acoustic analysis
convolutional neural network
classification
url https://doi.org/10.1002/lio2.259
work_keys_str_mv AT mariaepowell decodingphonationwithartificialintelligencedepaiproofofconcept
AT marcelinorodriguezcancio decodingphonationwithartificialintelligencedepaiproofofconcept
AT davidyoung decodingphonationwithartificialintelligencedepaiproofofconcept
AT williamnock decodingphonationwithartificialintelligencedepaiproofofconcept
AT beshoyabdelmessih decodingphonationwithartificialintelligencedepaiproofofconcept
AT amyzeller decodingphonationwithartificialintelligencedepaiproofofconcept
AT irvinperezmorales decodingphonationwithartificialintelligencedepaiproofofconcept
AT pengzhang decodingphonationwithartificialintelligencedepaiproofofconcept
AT cgaelyngarrett decodingphonationwithartificialintelligencedepaiproofofconcept
AT douglasschmidt decodingphonationwithartificialintelligencedepaiproofofconcept
AT juleswhite decodingphonationwithartificialintelligencedepaiproofofconcept
AT alexandergelbard decodingphonationwithartificialintelligencedepaiproofofconcept
_version_ 1724700088216846336