301

Beyond Invariance: Test-Time Label-Shift Adaptation for Distributions with "Spurious" Correlations

Neural Information Processing Systems (NeurIPS), 2022
Abstract

Changes in the data distribution at test time can have deleterious effects on the performance of predictive models p(yx)p(y|x). We consider situations where there are additional meta-data labels (such as group labels), denoted by zz, that can account for such changes in the distribution. In particular, we assume that the prior distribution p(y,z)p(y, z), which models the dependence between the class label yy and the "nuisance" factors zz, may change across domains, either due to a change in the correlation between these terms, or a change in one of their marginals. However, we assume that the generative model for features p(xy,z)p(x|y, z) is invariant across domains. We note that this corresponds to an expanded version of the widely used "label shift" assumption, where the labels now also include the nuisance factors zz. Based on this observation, we propose a test-time label shift correction that adapts to changes in the joint distribution p(y,z)p(y, z) using EM applied to unlabeled samples from the target domain distribution, pt(x)p_t(x). Importantly, we are able to avoid fitting a generative model p(xy,z)p(x|y,z), and merely need to reweight the outputs of a discriminative model ps(y,zx)p_s(y,z|x) trained on the source distribution. We evaluate our method, which we call "Test-Time Label-Shift Adaptation" (TTLSA), on several standard image and text datasets, as well as the CheXpert chest X-ray dataset, and show that it improves performance over methods that target invariance to changes in the distribution, as well as baseline empirical risk minimization methods. Code for reproducing experiments is available at https://github.com/nalzok/test-time-label-shift .

View on arXiv
Comments on this paper