432

Learning with Good Feature Representations in Bandits and in RL with a Generative Model

International Conference on Machine Learning (ICML), 2019
Abstract

The construction in the recent paper by Du et al. [2019] implies that searching for a near-optimal action in a bandit sometimes requires examining essentially all the actions, even if the learner is given linear features in Rd\mathbb R^d that approximate the rewards with a small uniform error. In this note we use the Kiefer-Wolfowitz theorem to show that by checking only a few actions, a learner can always find an action which is suboptimal with an error of at most O(εd)O(\varepsilon \sqrt{d}) where ε\varepsilon is the approximation error of the features. Thus, features are useful when the approximation error is small relative to the dimensionality of the features. The idea is applied to stochastic bandits and reinforcement learning with a generative model where the learner has access to dd-dimensional linear features that approximate the action-value functions for all policies to an accuracy of ε\varepsilon. For bandits we prove a bound on the regret of order dnlog(k)+εndlog(n)\sqrt{dn \log(k)} + \varepsilon n \sqrt{d} \log(n) with kk the number of actions and nn the horizon. For RL we show that approximate policy iteration can learn a policy that is optimal up to an additive error of order εd/(1γ)2\varepsilon \sqrt{d} / (1 - \gamma)^2 and using about d/(ε2(1γ)4)d / (\varepsilon^2(1-\gamma)^4) samples from the generative model.

View on arXiv
Comments on this paper