Clinical Prediction from Structural Brain MRI Scans: A Large-Scale Empirical Study
Multivariate pattern analysis (MVPA) methods have become an important tool in neuroimaging, revealing complex associations and yielding powerful prediction models. Despite methodological developments and novel application domains, there has been little effort to compile benchmark results that resear...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
Springer US,
2016-12-16T23:44:38Z.
|
Subjects: | |
Online Access: | Get fulltext |
Summary: | Multivariate pattern analysis (MVPA) methods have become an important tool in neuroimaging, revealing complex associations and yielding powerful prediction models. Despite methodological developments and novel application domains, there has been little effort to compile benchmark results that researchers can reference and compare against. This study takes a significant step in this direction. We employed three classes of state-of-the-art MVPA algorithms and common types of structural measurements from brain Magnetic Resonance Imaging (MRI) scans to predict an array of clinically relevant variables (diagnosis of Alzheimer's, schizophrenia, autism, and attention deficit and hyperactivity disorder; age, cerebrospinal fluid derived amyloid-β levels and mini-mental state exam score). We analyzed data from over 2,800 subjects, compiled from six publicly available datasets. The employed data and computational tools are freely distributed (https://www.nmr.mgh.harvard.edu/lab/mripredict), making this the largest, most comprehensive, reproducible benchmark image-based prediction experiment to date in structural neuroimaging. Finally, we make several observations regarding the factors that influence prediction performance and point to future research directions. Unsurprisingly, our results suggest that the biological footprint (effect size) has a dramatic influence on prediction performance. Though the choice of image measurement and MVPA algorithm can impact the result, there was no universally optimal selection. Intriguingly, the choice of algorithm seemed to be less critical than the choice of measurement type. Finally, our results showed that cross-validation estimates of performance, while generally optimistic, correlate well with generalization accuracy on a new dataset. BrightFocus Foundation (Alzheimer's Disease pilot grant (AHAF A2012333)) National Institutes of Health (U.S.) (K25 grant (NIBIB 1K25EB013649-01)) National Center for Research Resources (U.S.) (U24 RR021382) National Institutes of Health. National Institute for Biomedical Imaging and Bioengineering (R01EB006758) National Institute of Neurological Disorders and Stroke (U.S.) (R01 NS052585-01, 1R21NS072652-01, 1R01NS070963, R01NS083534) National Institutes of Health (U.S.) (Blueprint for Neuroscience Research (5U01-MH093765)) |
---|