Decoding phonation with artificial intelligence (DeP AI): Proof of concept
Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this...
Main Authors: | , , , , , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2019-06-01
|
Series: | Laryngoscope Investigative Otolaryngology |
Subjects: | |
Online Access: | https://doi.org/10.1002/lio2.259 |
id |
doaj-9a0bbc3973284e13a5f811f602615336 |
---|---|
record_format |
Article |
spelling |
doaj-9a0bbc3973284e13a5f811f6026153362020-11-25T02:59:57ZengWileyLaryngoscope Investigative Otolaryngology2378-80382019-06-014332833410.1002/lio2.259Decoding phonation with artificial intelligence (DeP AI): Proof of conceptMaria E. Powell0Marcelino Rodriguez Cancio1David Young2William Nock3Beshoy Abdelmessih4Amy Zeller5Irvin Perez Morales6Peng Zhang7C. Gaelyn Garrett8Douglas Schmidt9Jules White10Alexander Gelbard11Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Information Technology Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Center of Research in Computational and Numerical Methods in Engineering Central University Marta Abreu of Las Villas Santa Clara CubaDepartment of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Department of Electrical Engineering and Computer Science Vanderbilt University Nashville Tennessee U.S.A.Vanderbilt Bill Wilkerson Center for Otolaryngology Vanderbilt University Medical Center Nashville Tennessee U.S.A.Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this study is to provide a proof‐of‐concept that embedded data within human phonation can be accurately and efficiently decoded with deep learning neural network analysis to differentiate between normal and disordered voices. Methods Acoustic recordings from 10 vocally‐healthy speakers, as well as 70 patients with one of seven voice disorders (n = 10 per diagnosis), were acquired from a clinical database. Acoustic signals were converted into spectrograms and used to train a convolutional neural network developed with the Keras library. The network architecture was trained separately for each of the seven diagnostic categories. Binary classification tasks (ie, to classify normal vs. disordered) were performed for each of the seven diagnostic categories. All models were validated using the 10‐fold cross‐validation technique. Results Binary classification averaged accuracies ranged from 58% to 90%. Models were most accurate in their classification of adductor spasmodic dysphonia, unilateral vocal fold paralysis, vocal fold polyp, polypoid corditis, and recurrent respiratory papillomatosis. Despite a small sample size, these findings are consistent with previously published data utilizing deep neural networks for classification of voice disorders. Conclusion Promising preliminary results support further study of deep neural networks for clinical detection and diagnosis of human voice disorders. Current models should be optimized with a larger sample size. Levels of Evidence Level IIIhttps://doi.org/10.1002/lio2.259Voice disordersdetectionacoustic analysisconvolutional neural networkclassification |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Maria E. Powell Marcelino Rodriguez Cancio David Young William Nock Beshoy Abdelmessih Amy Zeller Irvin Perez Morales Peng Zhang C. Gaelyn Garrett Douglas Schmidt Jules White Alexander Gelbard |
spellingShingle |
Maria E. Powell Marcelino Rodriguez Cancio David Young William Nock Beshoy Abdelmessih Amy Zeller Irvin Perez Morales Peng Zhang C. Gaelyn Garrett Douglas Schmidt Jules White Alexander Gelbard Decoding phonation with artificial intelligence (DeP AI): Proof of concept Laryngoscope Investigative Otolaryngology Voice disorders detection acoustic analysis convolutional neural network classification |
author_facet |
Maria E. Powell Marcelino Rodriguez Cancio David Young William Nock Beshoy Abdelmessih Amy Zeller Irvin Perez Morales Peng Zhang C. Gaelyn Garrett Douglas Schmidt Jules White Alexander Gelbard |
author_sort |
Maria E. Powell |
title |
Decoding phonation with artificial intelligence (DeP AI): Proof of concept |
title_short |
Decoding phonation with artificial intelligence (DeP AI): Proof of concept |
title_full |
Decoding phonation with artificial intelligence (DeP AI): Proof of concept |
title_fullStr |
Decoding phonation with artificial intelligence (DeP AI): Proof of concept |
title_full_unstemmed |
Decoding phonation with artificial intelligence (DeP AI): Proof of concept |
title_sort |
decoding phonation with artificial intelligence (dep ai): proof of concept |
publisher |
Wiley |
series |
Laryngoscope Investigative Otolaryngology |
issn |
2378-8038 |
publishDate |
2019-06-01 |
description |
Objective Acoustic analysis of voice has the potential to expedite detection and diagnosis of voice disorders. Applying an image‐based, neural‐network approach to analyzing the acoustic signal may be an effective means for detecting and differentially diagnosing voice disorders. The purpose of this study is to provide a proof‐of‐concept that embedded data within human phonation can be accurately and efficiently decoded with deep learning neural network analysis to differentiate between normal and disordered voices. Methods Acoustic recordings from 10 vocally‐healthy speakers, as well as 70 patients with one of seven voice disorders (n = 10 per diagnosis), were acquired from a clinical database. Acoustic signals were converted into spectrograms and used to train a convolutional neural network developed with the Keras library. The network architecture was trained separately for each of the seven diagnostic categories. Binary classification tasks (ie, to classify normal vs. disordered) were performed for each of the seven diagnostic categories. All models were validated using the 10‐fold cross‐validation technique. Results Binary classification averaged accuracies ranged from 58% to 90%. Models were most accurate in their classification of adductor spasmodic dysphonia, unilateral vocal fold paralysis, vocal fold polyp, polypoid corditis, and recurrent respiratory papillomatosis. Despite a small sample size, these findings are consistent with previously published data utilizing deep neural networks for classification of voice disorders. Conclusion Promising preliminary results support further study of deep neural networks for clinical detection and diagnosis of human voice disorders. Current models should be optimized with a larger sample size. Levels of Evidence Level III |
topic |
Voice disorders detection acoustic analysis convolutional neural network classification |
url |
https://doi.org/10.1002/lio2.259 |
work_keys_str_mv |
AT mariaepowell decodingphonationwithartificialintelligencedepaiproofofconcept AT marcelinorodriguezcancio decodingphonationwithartificialintelligencedepaiproofofconcept AT davidyoung decodingphonationwithartificialintelligencedepaiproofofconcept AT williamnock decodingphonationwithartificialintelligencedepaiproofofconcept AT beshoyabdelmessih decodingphonationwithartificialintelligencedepaiproofofconcept AT amyzeller decodingphonationwithartificialintelligencedepaiproofofconcept AT irvinperezmorales decodingphonationwithartificialintelligencedepaiproofofconcept AT pengzhang decodingphonationwithartificialintelligencedepaiproofofconcept AT cgaelyngarrett decodingphonationwithartificialintelligencedepaiproofofconcept AT douglasschmidt decodingphonationwithartificialintelligencedepaiproofofconcept AT juleswhite decodingphonationwithartificialintelligencedepaiproofofconcept AT alexandergelbard decodingphonationwithartificialintelligencedepaiproofofconcept |
_version_ |
1724700088216846336 |