ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.12988
70
1
v1v2v3 (latest)

Why Ask One When You Can Ask kkk? Two-Stage Learning-to-Defer to the Top-kkk Experts

17 April 2025
Yannis Montreuil
Axel Carlier
Lai Xing Ng
Wei Tsang Ooi
ArXiv (abs)PDFHTML
Abstract

Learning-to-Defer (L2D) enables decision-making systems to improve reliability by selectively deferring uncertain predictions to more competent agents. However, most existing approaches focus exclusively on single-agent deferral, which is often inadequate in high-stakes scenarios that require collective expertise. We propose Top-kkk Learning-to-Defer, a generalization of the classical two-stage L2D framework that allocates each query to the kkk most confident agents instead of a single one. To further enhance flexibility and cost-efficiency, we introduce Top-k(x)k(x)k(x) Learning-to-Defer, an adaptive extension that learns the optimal number of agents to consult for each query, based on input complexity, agent competency distributions, and consultation costs. For both settings, we derive a novel surrogate loss and prove that it is Bayes-consistent and (R,G)(\mathcal{R}, \mathcal{G})(R,G)-consistent, ensuring convergence to the Bayes-optimal allocation. Notably, we show that the well-established model cascades paradigm arises as a restricted instance of our Top-kkk and Top-k(x)k(x)k(x) formulations. Extensive experiments across diverse benchmarks demonstrate the effectiveness of our framework on both classification and regression tasks.

View on arXiv
@article{montreuil2025_2504.12988,
  title={ Why Ask One When You Can Ask $k$? Two-Stage Learning-to-Defer to the Top-$k$ Experts },
  author={ Yannis Montreuil and Axel Carlier and Lai Xing Ng and Wei Tsang Ooi },
  journal={arXiv preprint arXiv:2504.12988},
  year={ 2025 }
}
Comments on this paper