240

Hardness of Maximum Likelihood Learning of DPPs

Annual Conference Computational Learning Theory (COLT), 2022
Main:1 Pages
10 Figures
Appendix:51 Pages
Abstract

Determinantal Point Processes (DPPs) are a widely used probabilistic model for negatively correlated sets. DPPs have been successfully employed in Machine Learning applications to select a diverse, yet representative subset of data. In seminal work on DPPs in Machine Learning, Kulesza conjectured in his PhD Thesis (2011) that the problem is NP-complete. The lack of a formal proof prompted Brunel, Moitra, Rigollet and Urschel (COLT 2017) to conjecture that, in opposition to Kulesza's conjecture, there exists a polynomial-time algorithm for computing a maximum-likelihood DPP. They also presented some preliminary evidence supporting their conjecture. In this work we prove Kulesza's conjecture. In fact, we prove the following stronger hardness of approximation result: even computing a (1O(1log9N))\left(1-O(\frac{1}{\log^9{N}})\right)-approximation to the maximum log-likelihood of a DPP on a ground set of NN elements is NP-complete. At the same time, we also obtain the first polynomial-time algorithm that achieves a nontrivial worst-case approximation to the optimal log-likelihood: the approximation factor is 1(1+o(1))logm\frac{1}{(1+o(1))\log{m}} unconditionally (for data sets that consist of mm subsets), and can be improved to 11+o(1)logN1-\frac{1+o(1)}{\log N} if all NN elements appear in a O(1/N)O(1/N)-fraction of the subsets. In terms of techniques, we reduce approximating the maximum log-likelihood of DPPs on a data set to solving a gap instance of a "vector coloring" problem on a hypergraph. Such a hypergraph is built on a bounded-degree graph construction of Bogdanov, Obata and Trevisan (FOCS 2002), and is further enhanced by the strong expanders of Alon and Capalbo (FOCS 2007) to serve our purposes.

View on arXiv
Comments on this paper