256

k-expertsk\texttt{-experts} -- Online Policies and Fundamental Limits

Main:8 Pages
8 Figures
Bibliography:3 Pages
3 Tables
Appendix:13 Pages
Abstract

This paper introduces and studies the k-expertsk\texttt{-experts} problem -- a generalization of the classic Prediction with Expert's Advice (i.e., the Experts\texttt{Experts}) problem. Unlike the Experts\texttt{Experts} problem, where the learner chooses exactly one expert, in this problem, the learner selects a subset of kk experts from a pool of NN experts at each round. The reward obtained by the learner at any round depends on the rewards of the selected experts. The k-expertsk\texttt{-experts} problem arises in many practical settings, including online ad placements, personalized news recommendations, and paging. Our primary goal is to design an online learning policy having a small regret. In this pursuit, we propose SAGE\texttt{SAGE} (Sa\textbf{Sa}mpled Hedge\textbf{ge}) - a framework for designing efficient online learning policies by leveraging statistical sampling techniques. We show that, for many related problems, SAGE\texttt{SAGE} improves upon the state-of-the-art bounds for regret and computational complexity. Furthermore, going beyond the notion of regret, we characterize the mistake bounds achievable by online learning policies for a class of stable loss functions. We conclude the paper by establishing a tight regret lower bound for a variant of the k-expertsk\texttt{-experts} problem and carrying out experiments with standard datasets.

View on arXiv
Comments on this paper