Information-Theoretic Inference of Common Ancestors
A directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2015-04-01
|
Series: | Entropy |
Subjects: | |
Online Access: | http://www.mdpi.com/1099-4300/17/4/2304 |
id |
doaj-cc622236f20540fa9ae7dca64c26488c |
---|---|
record_format |
Article |
spelling |
doaj-cc622236f20540fa9ae7dca64c26488c2020-11-24T21:51:54ZengMDPI AGEntropy1099-43002015-04-011742304232710.3390/e17042304e17042304Information-Theoretic Inference of Common AncestorsBastian Steudel0Nihat Ay1Max Planck Institute for Mathematics in the Sciences, Inselstraße 22, 04103 Leipzig, GermanyMax Planck Institute for Mathematics in the Sciences, Inselstraße 22, 04103 Leipzig, GermanyA directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents a given set of conditional independence relations. We are interested in properties of this class that can be derived from observations of a subsystem only. To this end, we prove an information-theoretic inequality that allows for the inference of common ancestors of observed parts in any DAG representing some unknown larger system. More explicitly, we show that a large amount of dependence in terms of mutual information among the observations implies the existence of a common ancestor that distributes this information. Within the causal interpretation of DAGs, our result can be seen as a quantitative extension of Reichenbach’s principle of common cause to more than two variables. Our conclusions are valid also for non-probabilistic observations, such as binary strings, since we state the proof for an axiomatized notion of “mutual information” that includes the stochastic as well as the algorithmic version.http://www.mdpi.com/1099-4300/17/4/2304information theorycommon cause principledirected acyclic graphsBayesian netscausalitymutual informationKolmogorov complexity |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Bastian Steudel Nihat Ay |
spellingShingle |
Bastian Steudel Nihat Ay Information-Theoretic Inference of Common Ancestors Entropy information theory common cause principle directed acyclic graphs Bayesian nets causality mutual information Kolmogorov complexity |
author_facet |
Bastian Steudel Nihat Ay |
author_sort |
Bastian Steudel |
title |
Information-Theoretic Inference of Common Ancestors |
title_short |
Information-Theoretic Inference of Common Ancestors |
title_full |
Information-Theoretic Inference of Common Ancestors |
title_fullStr |
Information-Theoretic Inference of Common Ancestors |
title_full_unstemmed |
Information-Theoretic Inference of Common Ancestors |
title_sort |
information-theoretic inference of common ancestors |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2015-04-01 |
description |
A directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents a given set of conditional independence relations. We are interested in properties of this class that can be derived from observations of a subsystem only. To this end, we prove an information-theoretic inequality that allows for the inference of common ancestors of observed parts in any DAG representing some unknown larger system. More explicitly, we show that a large amount of dependence in terms of mutual information among the observations implies the existence of a common ancestor that distributes this information. Within the causal interpretation of DAGs, our result can be seen as a quantitative extension of Reichenbach’s principle of common cause to more than two variables. Our conclusions are valid also for non-probabilistic observations, such as binary strings, since we state the proof for an axiomatized notion of “mutual information” that includes the stochastic as well as the algorithmic version. |
topic |
information theory common cause principle directed acyclic graphs Bayesian nets causality mutual information Kolmogorov complexity |
url |
http://www.mdpi.com/1099-4300/17/4/2304 |
work_keys_str_mv |
AT bastiansteudel informationtheoreticinferenceofcommonancestors AT nihatay informationtheoreticinferenceofcommonancestors |
_version_ |
1725877915904638976 |