Proofs as Explanations: Short Certificates for Reliable Predictions
- FAtt
We consider a model for explainable AI in which an explanation for a prediction consists of a subset of the training data (if it exists) such that all classifiers that make at most mistakes on predict . Such a set serves as a proof that indeed has label under the assumption that (1) the target function belongs to , and (2) the set contains at most corrupted points. For example, if and is the family of linear classifiers in , and if lies inside the convex hull of the positive data points in (and hence every consistent linear classifier labels as positive), then Carathéodory's theorem states that lies inside the convex hull of of those points. So, a set of size could be released as an explanation for a positive prediction, and would serve as a short proof of correctness of the prediction under the assumption of realizability.In this work, we consider this problem more generally, for general hypothesis classes and general values . We define the notion of the robust hollow star number of (which generalizes the standard hollow star number), and show that it precisely characterizes the worst-case size of the smallest certificate achievable, and analyze its size for natural classes. We also consider worst-case distributional bounds on certificate size, as well as distribution-dependent bounds that we show tightly control the sample size needed to get a certificate for any given test example. In particular, we define a notion of the certificate coefficient of an example with respect to a data distribution and target function , and prove matching upper and lower bounds on sample size as a function of , , and the VC dimension of .
View on arXiv