136
v1v2 (latest)

Towards a Learning Theory of Cause-Effect Inference

Abstract

We pose causal inference as the problem of learning to classify probability distributions. In particular, we assume access to a collection {(Si,li)}i=1n\{(S_i,l_i)\}_{i=1}^n, where each SiS_i is a sample drawn from the probability distribution of Xi×YiX_i \times Y_i, and lil_i is a binary label indicating whether "XiYiX_i \to Y_i" or "XiYiX_i \leftarrow Y_i". Given these data, we build a causal inference rule in two steps. First, we featurize each SiS_i using the kernel mean embedding associated with some characteristic kernel. Second, we train a binary classifier on such embeddings to distinguish between causal directions. We present generalization bounds showing the statistical consistency and learning rates of the proposed approach, and provide a simple implementation that achieves state-of-the-art cause-effect inference. Furthermore, we extend our ideas to infer causal relationships between more than two variables.

View on arXiv
Comments on this paper