17
3

On Submodular Contextual Bandits

Abstract

We consider the problem of contextual bandits where actions are subsets of a ground set and mean rewards are modeled by an unknown monotone submodular function that belongs to a class F\mathcal{F}. We allow time-varying matroid constraints to be placed on the feasible sets. Assuming access to an online regression oracle with regret Reg(F)\mathsf{Reg}(\mathcal{F}), our algorithm efficiently randomizes around local optima of estimated functions according to the Inverse Gap Weighting strategy. We show that cumulative regret of this procedure with time horizon nn scales as O(nReg(F))O(\sqrt{n \mathsf{Reg}(\mathcal{F})}) against a benchmark with a multiplicative factor 1/21/2. On the other hand, using the techniques of (Filmus and Ward 2014), we show that an ϵ\epsilon-Greedy procedure with local randomization attains regret of O(n2/3Reg(F)1/3)O(n^{2/3} \mathsf{Reg}(\mathcal{F})^{1/3}) against a stronger (1e1)(1-e^{-1}) benchmark.

View on arXiv
Comments on this paper