41

Normalizing Flow to Augmented Posterior: Conditional Density Estimation with Interpretable Dimension Reduction for High Dimensional Data

Cheng Zeng
George Michailidis
Hitoshi Iyatomi
Leo L Duan
Main:17 Pages
13 Figures
Bibliography:5 Pages
4 Tables
Appendix:3 Pages
Abstract

The conditional density characterizes the distribution of a response variable yy given other predictor xx, and plays a key role in many statistical tasks, including classification and outlier detection. Although there has been abundant work on the problem of Conditional Density Estimation (CDE) for a low-dimensional response in the presence of a high-dimensional predictor, little work has been done for a high-dimensional response such as images. The promising performance of normalizing flow (NF) neural networks in unconditional density estimation acts a motivating starting point. In this work, we extend NF neural networks when external xx is present. Specifically, they use the NF to parameterize a one-to-one transform between a high-dimensional yy and a latent zz that comprises two components \([z_P,z_N]\). The zPz_P component is a low-dimensional subvector obtained from the posterior distribution of an elementary predictive model for xx, such as logistic/linear regression. The zNz_N component is a high-dimensional independent Gaussian vector, which explains the variations in yy not or less related to xx. Unlike existing CDE methods, the proposed approach, coined Augmented Posterior CDE (AP-CDE), only requires a simple modification on the common normalizing flow framework, while significantly improving the interpretation of the latent component, since zPz_P represents a supervised dimension reduction. In image analytics applications, AP-CDE shows good separation of xx-related variations due to factors such as lighting condition and subject id, from the other random variations. Further, the experiments show that an unconditional NF neural network, based on an unsupervised model of zz, such as Gaussian mixture, fails to generate interpretable results.

View on arXiv
Comments on this paper