Empirical Inference Article 2012

Information-geometric approach to inferring causal directions

Web
Thumb ticker sm dominik 270 cut s
Empirical Inference
  • Research Scientist
Thumb ticker sm l1170153
Empirical Inference
  • Director
no image
Empirical Inference
Thumb ticker sm thumb kun zhang
Empirical Inference
Thumb ticker sm jakob zscheischler
Empirical Inference
no image
Empirical Inference

While conventional approaches to causal inference are mainly based on conditional (in)dependences, recent methods also account for the shape of (conditional) distributions. The idea is that the causal hypothesis “X causes Y” imposes that the marginal distribution PX and the conditional distribution PY|X represent independent mechanisms of nature. Recently it has been postulated that the shortest description of the joint distribution PX,Y should therefore be given by separate descriptions of PX and PY|X. Since description length in the sense of Kolmogorov complexity is uncomputable, practical implementations rely on other notions of independence. Here we define independence via orthogonality in information space. This way, we can explicitly describe the kind of dependence that occurs between PY and PX|Y making the causal hypothesis “Y causes X” implausible. Remarkably, this asymmetry between cause and effect becomes particularly simple if X and Y are deterministically related. We present an inference method that works in this case. We also discuss some theoretical results for the non-deterministic case although it is not clear how to employ them for a more general inference method.

Author(s): Janzing, D. and Mooij, J. and Zhang, K. and Lemeire, J. and Zscheischler, J. and Daniušis, P. and Steudel, B. and Schölkopf, B.
Links:
Journal: Artificial Intelligence
Volume: 182-183
Pages: 1-31
Year: 2012
Month: May
Day: 0
Bibtex Type: Article (article)
DOI: 10.1016/j.artint.2012.01.002
Digital: 0
Electronic Archiving: grant_archive

BibTex

@article{JanzingMZLZDSS2012,
  title = {Information-geometric approach to inferring causal directions},
  journal = {Artificial Intelligence},
  abstract = {While conventional approaches to causal inference are mainly based on conditional (in)dependences, recent methods also account for the shape of (conditional) distributions. The idea is that the causal hypothesis “X causes Y” imposes that the marginal distribution PX and the conditional distribution PY|X represent independent mechanisms of nature. Recently it has been postulated that the shortest description of the joint distribution PX,Y should therefore be given by separate descriptions of PX and PY|X. Since description length in the sense of Kolmogorov complexity is uncomputable, practical implementations rely on other notions of independence. Here we define independence via orthogonality in information space. This way, we can explicitly describe the kind of dependence that occurs between PY and PX|Y making the causal hypothesis “Y causes X” implausible. Remarkably, this asymmetry between cause and effect becomes particularly simple if X and Y are deterministically related. We present an inference method that works in this case. We also discuss some theoretical results for the non-deterministic case although it is not clear how to employ them for a more general inference method.},
  volume = {182-183},
  pages = {1-31},
  month = may,
  year = {2012},
  slug = {janzingmzlzdss2012},
  author = {Janzing, D. and Mooij, J. and Zhang, K. and Lemeire, J. and Zscheischler, J. and Daniušis, P. and Steudel, B. and Sch{\"o}lkopf, B.},
  month_numeric = {5}
}