68
1
v1v2v3 (latest)

Why Ask One When You Can Ask kk? Two-Stage Learning-to-Defer to the Top-kk Experts

Abstract

Although existing Learning-to-Defer (L2D) frameworks support multiple experts, they allocate each query to a single expert, limiting their ability to leverage collective expertise in complex decision-making scenarios. To address this, we introduce the first framework for Top-kk Learning-to-Defer, enabling systems to defer each query to the kk most cost-effective experts. Our formulation strictly generalizes classical two-stage L2D by supporting multi-expert deferral-a capability absent in prior work. We further propose Top-k(x)k(x) Learning-to-Defer, an adaptive extension that learns the optimal number of experts per query based on input complexity, expert quality, and consultation cost. We introduce a novel surrogate loss that is Bayes-consistent, (R,G)(\mathcal{R}, \mathcal{G})-consistent, and independent of the cardinality parameter kk, enabling efficient reuse across different values of kk. We show that classical model cascades arise as a special case of our method, situating our framework as a strict generalization of both selective deferral and cascaded inference. Experiments on classification and regression demonstrate that Top-kk and Top-k(x)k(x) yield improved accuracy--cost trade-offs, establishing a new direction for multi-expert deferral in Learning-to-Defer.

View on arXiv
@article{montreuil2025_2504.12988,
  title={ Why Ask One When You Can Ask $k$? Two-Stage Learning-to-Defer to the Top-$k$ Experts },
  author={ Yannis Montreuil and Axel Carlier and Lai Xing Ng and Wei Tsang Ooi },
  journal={arXiv preprint arXiv:2504.12988},
  year={ 2025 }
}
Comments on this paper