700

Learning Causal Semantic Representation for Out-of-Distribution Prediction

Abstract

Conventional supervised learning methods, especially deep ones, are found to be sensitive to out-of-distribution (OOD) examples, largely because the learned representation mixes the semantic factor with the variation factor due to their domain-specific correlation, while only the semantic factor causes the output. To address the problem, we propose a Causal Semantic Generative model (CSG) based on a causal thought so that the two factors are modeled separately, and develop methods to learn it on a single training domain and predict in a test domain without (OOD generalization) or with unsupervised data (domain adaptation). We prove that under proper conditions, CSG identifies the semantic factor by learning from training data, and this semantic identification guarantees the boundedness of OOD generalization error and the success of adaptation. The methods and theory are built on the invariance principle of causal generative mechanisms, which is fundamental and general. The methods are based on variational Bayes, with a novel design for both efficient learning and easy prediction. Empirical study demonstrates the improved test accuracy for both OOD generalization and domain adaptation.

View on arXiv
Comments on this paper