11
1

Mediated Uncoupled Learning: Learning Functions without Direct Input-output Correspondences

Abstract

Ordinary supervised learning is useful when we have paired training data of input XX and output YY. However, such paired data can be difficult to collect in practice. In this paper, we consider the task of predicting YY from XX when we have no paired data of them, but we have two separate, independent datasets of XX and YY each observed with some mediating variable UU, that is, we have two datasets SX={(Xi,Ui)}S_X = \{(X_i, U_i)\} and SY={(Uj,Yj)}S_Y = \{(U'_j, Y'_j)\}. A naive approach is to predict UU from XX using SXS_X and then YY from UU using SYS_Y, but we show that this is not statistically consistent. Moreover, predicting UU can be more difficult than predicting YY in practice, e.g., when UU has higher dimensionality. To circumvent the difficulty, we propose a new method that avoids predicting UU but directly learns Y=f(X)Y = f(X) by training f(X)f(X) with SXS_{X} to predict h(U)h(U) which is trained with SYS_{Y} to approximate YY. We prove statistical consistency and error bounds of our method and experimentally confirm its practical usefulness.

View on arXiv
Comments on this paper