Outlier Robust Multivariate Polynomial Regression
We study the problem of robust multivariate polynomial regression: let be an unknown -variate polynomial of degree at most in each variable. We are given as input a set of random samples that are noisy versions of . More precisely, each is sampled independently from some distribution on , and for each independently, is arbitrary (i.e., an outlier) with probability at most , and otherwise satisfies . The goal is to output a polynomial , of degree at most in each variable, within an -distance of at most from . Kane, Karmalkar, and Price [FOCS'17] solved this problem for . We generalize their results to the -variate setting, showing an algorithm that achieves a sample complexity of , where the hidden constant depends on , if is the -dimensional Chebyshev distribution. The sample complexity is , if the samples are drawn from the uniform distribution instead. The approximation error is guaranteed to be at most , and the run-time depends on . In the setting where each and are known up to bits of precision, the run-time's dependence on is linear. We also show that our sample complexities are optimal in terms of . Furthermore, we show that it is possible to have the run-time be independent of , at the cost of a higher sample complexity.
View on arXiv