Contextual Bandits with Cross-learning

In the classical contextual bandits problem, in each round , a learner observes some context , chooses some action to perform, and receives some reward . We consider the variant of this problem where in addition to receiving the reward , the learner also learns the values of for all other contexts ; i.e., the rewards that would have been achieved by performing that action under different contexts. This variant arises in several strategic settings, such as learning how to bid in non-truthful repeated auctions (in this setting the context is the decision maker's private valuation for each auction). We call this problem the contextual bandits problem with cross-learning. The best algorithms for the classical contextual bandits problem achieve regret against all stationary policies, where is the number of contexts, the number of actions, and the number of rounds. We demonstrate algorithms for the contextual bandits problem with cross-learning that remove the dependence on and achieve regret (when contexts are stochastic with known distribution), (when contexts are stochastic with unknown distribution), and (when contexts are adversarial but rewards are stochastic).
View on arXiv