Information-Theoretic Inference of Common Ancestors

A directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents...

Full description

Bibliographic Details
Main Authors: Bastian Steudel, Nihat Ay
Format: Article
Language:English
Published: MDPI AG 2015-04-01
Series:Entropy
Subjects:
Online Access:http://www.mdpi.com/1099-4300/17/4/2304
id doaj-cc622236f20540fa9ae7dca64c26488c
record_format Article
spelling doaj-cc622236f20540fa9ae7dca64c26488c2020-11-24T21:51:54ZengMDPI AGEntropy1099-43002015-04-011742304232710.3390/e17042304e17042304Information-Theoretic Inference of Common AncestorsBastian Steudel0Nihat Ay1Max Planck Institute for Mathematics in the Sciences, Inselstraße 22, 04103 Leipzig, GermanyMax Planck Institute for Mathematics in the Sciences, Inselstraße 22, 04103 Leipzig, GermanyA directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents a given set of conditional independence relations. We are interested in properties of this class that can be derived from observations of a subsystem only. To this end, we prove an information-theoretic inequality that allows for the inference of common ancestors of observed parts in any DAG representing some unknown larger system. More explicitly, we show that a large amount of dependence in terms of mutual information among the observations implies the existence of a common ancestor that distributes this information. Within the causal interpretation of DAGs, our result can be seen as a quantitative extension of Reichenbach’s principle of common cause to more than two variables. Our conclusions are valid also for non-probabilistic observations, such as binary strings, since we state the proof for an axiomatized notion of “mutual information” that includes the stochastic as well as the algorithmic version.http://www.mdpi.com/1099-4300/17/4/2304information theorycommon cause principledirected acyclic graphsBayesian netscausalitymutual informationKolmogorov complexity
collection DOAJ
language English
format Article
sources DOAJ
author Bastian Steudel
Nihat Ay
spellingShingle Bastian Steudel
Nihat Ay
Information-Theoretic Inference of Common Ancestors
Entropy
information theory
common cause principle
directed acyclic graphs
Bayesian nets
causality
mutual information
Kolmogorov complexity
author_facet Bastian Steudel
Nihat Ay
author_sort Bastian Steudel
title Information-Theoretic Inference of Common Ancestors
title_short Information-Theoretic Inference of Common Ancestors
title_full Information-Theoretic Inference of Common Ancestors
title_fullStr Information-Theoretic Inference of Common Ancestors
title_full_unstemmed Information-Theoretic Inference of Common Ancestors
title_sort information-theoretic inference of common ancestors
publisher MDPI AG
series Entropy
issn 1099-4300
publishDate 2015-04-01
description A directed acyclic graph (DAG) partially represents the conditional independence structure among observations of a system if the local Markov condition holds, that is if every variable is independent of its non-descendants given its parents. In general, there is a whole class of DAGs that represents a given set of conditional independence relations. We are interested in properties of this class that can be derived from observations of a subsystem only. To this end, we prove an information-theoretic inequality that allows for the inference of common ancestors of observed parts in any DAG representing some unknown larger system. More explicitly, we show that a large amount of dependence in terms of mutual information among the observations implies the existence of a common ancestor that distributes this information. Within the causal interpretation of DAGs, our result can be seen as a quantitative extension of Reichenbach’s principle of common cause to more than two variables. Our conclusions are valid also for non-probabilistic observations, such as binary strings, since we state the proof for an axiomatized notion of “mutual information” that includes the stochastic as well as the algorithmic version.
topic information theory
common cause principle
directed acyclic graphs
Bayesian nets
causality
mutual information
Kolmogorov complexity
url http://www.mdpi.com/1099-4300/17/4/2304
work_keys_str_mv AT bastiansteudel informationtheoreticinferenceofcommonancestors
AT nihatay informationtheoreticinferenceofcommonancestors
_version_ 1725877915904638976