Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers

Recent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to com...

Full description

Bibliographic Details
Main Authors: Kathleen C. Fraser, Kristina Lundholm Fors, Marie Eckerström, Fredrik Öhman, Dimitrios Kokkinakis
Format: Article
Language:English
Published: Frontiers Media S.A. 2019-08-01
Series:Frontiers in Aging Neuroscience
Subjects:
Online Access:https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/full
id doaj-9b609e561b214837bd7fe8c8ccf7f2e1
record_format Article
spelling doaj-9b609e561b214837bd7fe8c8ccf7f2e12020-11-25T01:52:33ZengFrontiers Media S.A.Frontiers in Aging Neuroscience1663-43652019-08-011110.3389/fnagi.2019.00205457880Predicting MCI Status From Multimodal Language Data Using Cascaded ClassifiersKathleen C. Fraser0Kathleen C. Fraser1Kristina Lundholm Fors2Marie Eckerström3Fredrik Öhman4Fredrik Öhman5Dimitrios Kokkinakis6Digital Technologies Research Centre, National Research Council Canada, Ottawa, ON, CanadaDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenInstitute of Neuroscience and Physiology, Sahlgrenska Academy, Gothenburg, SwedenInstitute of Neuroscience and Physiology, Sahlgrenska Academy, Gothenburg, SwedenWallenberg Centre for Molecular and Translational Medicine, University of Gothenburg, Gothenburg, SwedenDepartment of Swedish, University of Gothenburg, Gothenburg, SwedenRecent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to combine data from multiple language tasks. A cohort of 26 MCI participants and 29 healthy controls completed three language tasks: picture description, reading silently, and reading aloud. Information from each task is captured through different modes (audio, text, eye-tracking, and comprehension questions). Features are extracted from each mode, and used to train a series of cascaded classifiers which output predictions at the level of features, modes, tasks, and finally at the overall session level. The best classification result is achieved through combining the data at the task level (AUC = 0.88, accuracy = 0.83). This outperforms a classifier trained on neuropsychological test scores (AUC = 0.75, accuracy = 0.65) as well as the “early fusion” approach to multimodal classification (AUC = 0.79, accuracy = 0.70). By combining the predictions from the multimodal language classifier and the neuropsychological classifier, this result can be further improved to AUC = 0.90 and accuracy = 0.84. In a correlation analysis, language classifier predictions are found to be moderately correlated (ρ = 0.42) with participant scores on the Rey Auditory Verbal Learning Test (RAVLT). The cascaded approach for multimodal classification improves both system performance and interpretability. This modular architecture can be easily generalized to incorporate different types of classifiers as well as other heterogeneous sources of data (imaging, metabolic, etc.).https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/fullmild cognitive impairmentlanguagespeecheye-trackingmachine learningmultimodal
collection DOAJ
language English
format Article
sources DOAJ
author Kathleen C. Fraser
Kathleen C. Fraser
Kristina Lundholm Fors
Marie Eckerström
Fredrik Öhman
Fredrik Öhman
Dimitrios Kokkinakis
spellingShingle Kathleen C. Fraser
Kathleen C. Fraser
Kristina Lundholm Fors
Marie Eckerström
Fredrik Öhman
Fredrik Öhman
Dimitrios Kokkinakis
Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
Frontiers in Aging Neuroscience
mild cognitive impairment
language
speech
eye-tracking
machine learning
multimodal
author_facet Kathleen C. Fraser
Kathleen C. Fraser
Kristina Lundholm Fors
Marie Eckerström
Fredrik Öhman
Fredrik Öhman
Dimitrios Kokkinakis
author_sort Kathleen C. Fraser
title Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
title_short Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
title_full Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
title_fullStr Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
title_full_unstemmed Predicting MCI Status From Multimodal Language Data Using Cascaded Classifiers
title_sort predicting mci status from multimodal language data using cascaded classifiers
publisher Frontiers Media S.A.
series Frontiers in Aging Neuroscience
issn 1663-4365
publishDate 2019-08-01
description Recent work has indicated the potential utility of automated language analysis for the detection of mild cognitive impairment (MCI). Most studies combining language processing and machine learning for the prediction of MCI focus on a single language task; here, we consider a cascaded approach to combine data from multiple language tasks. A cohort of 26 MCI participants and 29 healthy controls completed three language tasks: picture description, reading silently, and reading aloud. Information from each task is captured through different modes (audio, text, eye-tracking, and comprehension questions). Features are extracted from each mode, and used to train a series of cascaded classifiers which output predictions at the level of features, modes, tasks, and finally at the overall session level. The best classification result is achieved through combining the data at the task level (AUC = 0.88, accuracy = 0.83). This outperforms a classifier trained on neuropsychological test scores (AUC = 0.75, accuracy = 0.65) as well as the “early fusion” approach to multimodal classification (AUC = 0.79, accuracy = 0.70). By combining the predictions from the multimodal language classifier and the neuropsychological classifier, this result can be further improved to AUC = 0.90 and accuracy = 0.84. In a correlation analysis, language classifier predictions are found to be moderately correlated (ρ = 0.42) with participant scores on the Rey Auditory Verbal Learning Test (RAVLT). The cascaded approach for multimodal classification improves both system performance and interpretability. This modular architecture can be easily generalized to incorporate different types of classifiers as well as other heterogeneous sources of data (imaging, metabolic, etc.).
topic mild cognitive impairment
language
speech
eye-tracking
machine learning
multimodal
url https://www.frontiersin.org/article/10.3389/fnagi.2019.00205/full
work_keys_str_mv AT kathleencfraser predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT kathleencfraser predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT kristinalundholmfors predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT marieeckerstrom predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT fredrikohman predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT fredrikohman predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
AT dimitrioskokkinakis predictingmcistatusfrommultimodallanguagedatausingcascadedclassifiers
_version_ 1724994534152077312