150
v1v2 (latest)

Impossibility of Depth Reduction in Explainable Clustering

Main:12 Pages
2 Figures
Bibliography:3 Pages
Abstract

Over the last few years Explainable Clustering has gathered a lot of attention. Dasgupta et al. [ICML'20] initiated the study of explainable kk-means and kk-median clustering problems where the explanation is captured by a threshold decision tree which partitions the space at each node using axis parallel hyperplanes. Recently, Laber et al. [Pattern Recognition'23] made a case to consider the depth of the decision tree as an additional complexity measure of interest.In this work, we prove that even when the input points are in the Euclidean plane, then any depth reduction in the explanation incurs unbounded loss in the kk-means and kk-median cost. Formally, we show that there exists a data set XR2X\subseteq \mathbb{R}^2, for which there is a decision tree of depth k1k-1 whose kk-means/kk-median cost matches the optimal clustering cost of XX, but every decision tree of depth less than k1k-1 has unbounded cost w.r.t. the optimal cost of clustering. We extend our results to the kk-center objective as well, albeit with weaker guarantees.

View on arXiv
Comments on this paper