Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
Recent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to com...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Frontiers Media S.A.
2019-08-01
|
Series: | Frontiers in Aging Neuroscience |
Subjects: | |
Online Access: | https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/full |
id |
doaj-9b609e561b214837bd7fe8c8ccf7f2e1 |
---|---|
record_format |
Article |
spelling |
doaj-9b609e561b214837bd7fe8c8ccf7f2e12020-11-25T01:52:33ZengFrontiers Media S.A.Frontiers in Aging Neuroscience1663-43652019-08-011110.3389/fnagi.2019.00205457880Predicting MCI Status From Multimodal Language Data Using Cascaded ClassifiersKathleen C. Fraser0Kathleen C. Fraser1Kristina Lundholm Fors2Marie Eckerström3Fredrik Öhman4Fredrik Öhman5Dimitrios Kokkinakis6Digital Technologies Research Centre, National Research Council Canada, Ottawa, ON, CanadaDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenInstitute of Neuroscience and Physiology, Sahlgrenska Academy, Gothenburg, SwedenInstitute of Neuroscience and Physiology, Sahlgrenska Academy, Gothenburg, SwedenWallenberg Centre for Molecular and Translational Medicine, University of Gothenburg, Gothenburg, SwedenDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenRecent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to combine data from multiple language tasks. A cohort of 26 MCI participants and 29 healthy controls completed three language tasks: picture description, reading silently, and reading aloud. Information from each task is captured through different modes (audio, text, eye-tracking, and comprehension questions). Features are extracted from each mode, and used to train a series of cascaded classifiers which output predictions at the level of features, modes, tasks, and finally at the overall session level. The best classification result is achieved through combining the data at the task level (AUC = 0.88, accuracy = 0.83). This outperforms a classifier trained on neuropsychological test scores (AUC = 0.75, accuracy = 0.65) as well as the “early fusion” approach to multimodal classification (AUC = 0.79, accuracy = 0.70). By combining the predictions from the multimodal language classifier and the neuropsychological classifier, this result can be further improved to AUC = 0.90 and accuracy = 0.84. In a correlation analysis, language classifier predictions are found to be moderately correlated (ρ = 0.42) with participant scores on the Rey Auditory Verbal Learning Test (RAVLT). The cascaded approach for multimodal classification improves both system performance and interpretability. This modular architecture can be easily generalized to incorporate different types of classifiers as well as other heterogeneous sources of data (imaging, metabolic, etc.).https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/fullmild cognitive impairmentlanguagespeecheye-trackingmachine learningmultimodal |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Kathleen C. Fraser Kathleen C. Fraser Kristina Lundholm Fors Marie Eckerström Fredrik Öhman Fredrik Öhman Dimitrios Kokkinakis |
spellingShingle |
Kathleen C. Fraser Kathleen C. Fraser Kristina Lundholm Fors Marie Eckerström Fredrik Öhman Fredrik Öhman Dimitrios Kokkinakis Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers Frontiers in Aging Neuroscience mild cognitive impairment language speech eye-tracking machine learning multimodal |
author_facet |
Kathleen C. Fraser Kathleen C. Fraser Kristina Lundholm Fors Marie Eckerström Fredrik Öhman Fredrik Öhman Dimitrios Kokkinakis |
author_sort |
Kathleen C. Fraser |
title |
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers |
title_short |
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers |
title_full |
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers |
title_fullStr |
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers |
title_full_unstemmed |
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers |
title_sort |
predicting mci status from multimodal language data using cascaded classifiers |
publisher |
Frontiers Media S.A. |
series |
Frontiers in Aging Neuroscience |
issn |
1663-4365 |
publishDate |
2019-08-01 |
description |
Recent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to combine data from multiple language tasks. A cohort of 26 MCI participants and 29 healthy controls completed three language tasks: picture description, reading silently, and reading aloud. Information from each task is captured through different modes (audio, text, eye-tracking, and comprehension questions). Features are extracted from each mode, and used to train a series of cascaded classifiers which output predictions at the level of features, modes, tasks, and finally at the overall session level. The best classification result is achieved through combining the data at the task level (AUC = 0.88, accuracy = 0.83). This outperforms a classifier trained on neuropsychological test scores (AUC = 0.75, accuracy = 0.65) as well as the “early fusion” approach to multimodal classification (AUC = 0.79, accuracy = 0.70). By combining the predictions from the multimodal language classifier and the neuropsychological classifier, this result can be further improved to AUC = 0.90 and accuracy = 0.84. In a correlation analysis, language classifier predictions are found to be moderately correlated (ρ = 0.42) with participant scores on the Rey Auditory Verbal Learning Test (RAVLT). The cascaded approach for multimodal classification improves both system performance and interpretability. This modular architecture can be easily generalized to incorporate different types of classifiers as well as other heterogeneous sources of data (imaging, metabolic, etc.). |
topic |
mild cognitive impairment language speech eye-tracking machine learning multimodal |
url |
https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/full |
work_keys_str_mv |
AT kathleencfraser predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT kathleencfraser predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT kristinalundholmfors predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT marieeckerstrom predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT fredrikohman predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT fredrikohman predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers AT dimitrioskokkinakis predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers |
_version_ |
1724994534152077312 |