A Review of the Application of Information Theory to Clinical Diagnostic Testing
The fundamental information theory functions of entropy, relative entropy, and mutual information are directly applicable to clinical diagnostic testing. This is a consequence of the fact that an individual’s disease state and diagnostic test result are random variables. In this paper, we...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2020-01-01
|
Series: | Entropy |
Subjects: | |
Online Access: | https://www.mdpi.com/1099-4300/22/1/97 |
id |
doaj-77317938b56340acbe13223c06c5f5bb |
---|---|
record_format |
Article |
spelling |
doaj-77317938b56340acbe13223c06c5f5bb2020-11-25T00:30:35ZengMDPI AGEntropy1099-43002020-01-012219710.3390/e22010097e22010097A Review of the Application of Information Theory to Clinical Diagnostic TestingWilliam A. Benish0Department of Internal Medicine, Case Western Reserve University, Cleveland, OH 44106, USAThe fundamental information theory functions of entropy, relative entropy, and mutual information are directly applicable to clinical diagnostic testing. This is a consequence of the fact that an individual’s disease state and diagnostic test result are random variables. In this paper, we review the application of information theory to the quantification of diagnostic uncertainty, diagnostic information, and diagnostic test performance. An advantage of information theory functions over more established test performance measures is that they can be used when multiple disease states are under consideration as well as when the diagnostic test can yield multiple or continuous results. Since more than one diagnostic test is often required to help determine a patient’s disease state, we also discuss the application of the theory to situations in which more than one diagnostic test is used. The total diagnostic information provided by two or more tests can be partitioned into meaningful components.https://www.mdpi.com/1099-4300/22/1/97entropyinformation theorymultiple diagnostic testsmutual informationrelative entropy |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
William A. Benish |
spellingShingle |
William A. Benish A Review of the Application of Information Theory to Clinical Diagnostic Testing Entropy entropy information theory multiple diagnostic tests mutual information relative entropy |
author_facet |
William A. Benish |
author_sort |
William A. Benish |
title |
A Review of the Application of Information Theory to Clinical Diagnostic Testing |
title_short |
A Review of the Application of Information Theory to Clinical Diagnostic Testing |
title_full |
A Review of the Application of Information Theory to Clinical Diagnostic Testing |
title_fullStr |
A Review of the Application of Information Theory to Clinical Diagnostic Testing |
title_full_unstemmed |
A Review of the Application of Information Theory to Clinical Diagnostic Testing |
title_sort |
review of the application of information theory to clinical diagnostic testing |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2020-01-01 |
description |
The fundamental information theory functions of entropy, relative entropy, and mutual information are directly applicable to clinical diagnostic testing. This is a consequence of the fact that an individual’s disease state and diagnostic test result are random variables. In this paper, we review the application of information theory to the quantification of diagnostic uncertainty, diagnostic information, and diagnostic test performance. An advantage of information theory functions over more established test performance measures is that they can be used when multiple disease states are under consideration as well as when the diagnostic test can yield multiple or continuous results. Since more than one diagnostic test is often required to help determine a patient’s disease state, we also discuss the application of the theory to situations in which more than one diagnostic test is used. The total diagnostic information provided by two or more tests can be partitioned into meaningful components. |
topic |
entropy information theory multiple diagnostic tests mutual information relative entropy |
url |
https://www.mdpi.com/1099-4300/22/1/97 |
work_keys_str_mv |
AT williamabenish areviewoftheapplicationofinformationtheorytoclinicaldiagnostictesting AT williamabenish reviewoftheapplicationofinformationtheorytoclinicaldiagnostictesting |
_version_ |
1725326056377810944 |