13

Detecting Unobserved Confounders: A Kernelized Regression Approach

Yikai Chen
Yunxin Mao
Chunyuan Zheng
Hao Zou
Shanzhi Gu
Shixuan Liu
Yang Shi
Wenjing Yang
Kun Kuang
Haotian Wang
Main:7 Pages
3 Figures
Bibliography:2 Pages
2 Tables
Appendix:5 Pages
Abstract

Detecting unobserved confounders is crucial for reliable causal inference in observational studies. Existing methods require either linearity assumptions or multiple heterogeneous environments, limiting applicability to nonlinear single-environment settings. To bridge this gap, we propose Kernel Regression Confounder Detection (KRCD), a novel method for detecting unobserved confounding in nonlinear observational data under single-environment conditions. KRCD leverages reproducing kernel Hilbert spaces to model complex dependencies. By comparing standard and higherorder kernel regressions, we derive a test statistic whose significant deviation from zero indicates unobserved confounding. Theoretically, we prove two key results: First, in infinite samples, regression coefficients coincide if and only if no unobserved confounders exist. Second, finite-sample differences converge to zero-mean Gaussian distributions with tractable variance. Extensive experiments on synthetic benchmarks and the Twins dataset demonstrate that KRCD not only outperforms existing baselines but also achieves superior computational efficiency.

View on arXiv
Comments on this paper