Dimension reduction as an optimization problem over a set of generalized functions

Classical dimension reduction problem can be loosely formulated as a problem of finding a -dimensional affine subspace of onto which data points can be projected without loss of valuable information. We reformulate this problem in the language of tempered distributions, i.e. as a problem of approximating an empirical probability density function , where is an -dimensional Dirac delta function, by another tempered distribution whose density is supported in some -dimensional subspace. Thus, our problem is reduced to the minimization of a certain loss function measuring the distance from to over a pertinent set of generalized functions, denoted . Another classical problem of data analysis is the sufficient dimension reduction problem. We show that it can be reduced to the following problem: given a function and a probability density function , find a function of the form that minimizes the loss . We first show that search spaces of the latter two problems are in one-to-one correspondence which is defined by the Fourier transform. We introduce a nonnegative penalty function and a set of ordinary functions in such a way that `approximates' the space when . Then we present an algorithm for minimization of , based on the idea of two-step iterative computation.
View on arXiv