34
0

Beyond Task Diversity: Provable Representation Transfer for Sequential Multi-Task Linear Bandits

Abstract

We study lifelong learning in linear bandits, where a learner interacts with a sequence of linear bandit tasks whose parameters lie in an mm-dimensional subspace of Rd\mathbb{R}^d, thereby sharing a low-rank representation. Current literature typically assumes that the tasks are diverse, i.e., their parameters uniformly span the mm-dimensional subspace. This assumption allows the low-rank representation to be learned before all tasks are revealed, which can be unrealistic in real-world applications. In this work, we present the first nontrivial result for sequential multi-task linear bandits without the task diversity assumption. We develop an algorithm that efficiently learns and transfers low-rank representations. When facing NN tasks, each played over τ\tau rounds, our algorithm achieves a regret guarantee of O~(Nmτ+N23τ23dm13+Nd2+τmd)\tilde{O}\big (Nm \sqrt{\tau} + N^{\frac{2}{3}} \tau^{\frac{2}{3}} d m^{\frac13} + Nd^2 + \tau m d \big) under the ellipsoid action set assumption. This result can significantly improve upon the baseline of O~(Ndτ)\tilde{O} \left (Nd \sqrt{\tau}\right) that does not leverage the low-rank structure when the number of tasks NN is sufficiently large and mdm \ll d. We also demonstrate empirically on synthetic data that our algorithm outperforms baseline algorithms, which rely on the task diversity assumption.

View on arXiv
Comments on this paper