Analysis validation has been neglected in the Age of Reproducibility.

Increasingly complex statistical models are being used for the analysis of biological data. Recent commentary has focused on the ability to compute the same outcome for a given dataset (reproducibility). We argue that a reproducible statistical analysis is not necessarily valid because of unique pat...

Full description

Bibliographic Details
Main Authors: Kathleen E Lotterhos, Jason H Moore, Ann E Stapleton
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2018-12-01
Series:PLoS Biology
Online Access:https://doi.org/10.1371/journal.pbio.3000070
Description
Summary:Increasingly complex statistical models are being used for the analysis of biological data. Recent commentary has focused on the ability to compute the same outcome for a given dataset (reproducibility). We argue that a reproducible statistical analysis is not necessarily valid because of unique patterns of nonindependence in every biological dataset. We advocate that analyses should be evaluated with known-truth simulations that capture biological reality, a process we call "analysis validation." We review the process of validation and suggest criteria that a validation project should meet. We find that different fields of science have historically failed to meet all criteria, and we suggest ways to implement meaningful validation in training and practice.
ISSN:1544-9173
1545-7885