Why Ask One When You Can Ask ? Two-Stage Learning-to-Defer to the Top- Experts

Although existing Learning-to-Defer (L2D) frameworks support multiple experts, they allocate each query to a single expert, limiting their ability to leverage collective expertise in complex decision-making scenarios. To address this, we introduce the first framework for Top- Learning-to-Defer, enabling systems to defer each query to the most cost-effective experts. Our formulation strictly generalizes classical two-stage L2D by supporting multi-expert deferral-a capability absent in prior work. We further propose Top- Learning-to-Defer, an adaptive extension that learns the optimal number of experts per query based on input complexity, expert quality, and consultation cost. We introduce a novel surrogate loss that is Bayes-consistent, -consistent, and independent of the cardinality parameter , enabling efficient reuse across different values of . We show that classical model cascades arise as a special case of our method, situating our framework as a strict generalization of both selective deferral and cascaded inference. Experiments on classification and regression demonstrate that Top- and Top- yield improved accuracy--cost trade-offs, establishing a new direction for multi-expert deferral in Learning-to-Defer.
View on arXiv@article{montreuil2025_2504.12988, title={ Why Ask One When You Can Ask $k$? Two-Stage Learning-to-Defer to the Top-$k$ Experts }, author={ Yannis Montreuil and Axel Carlier and Lai Xing Ng and Wei Tsang Ooi }, journal={arXiv preprint arXiv:2504.12988}, year={ 2025 } }