577

Efficient Exploration and Discriminative World Model Learning with an Object-Centric Abstraction

International Conference on Learning Representations (ICLR), 2024
Main:10 Pages
38 Figures
Bibliography:6 Pages
10 Tables
Appendix:22 Pages
Abstract

In the face of difficult exploration problems in reinforcement learning, we study whether giving an agent an object-centric mapping (describing a set of items and their attributes) allow for more efficient learning. We found this problem is best solved hierarchically by modelling items at a higher level of state abstraction to pixels, and attribute change at a higher level of temporal abstraction to primitive actions. This abstraction simplifies the transition dynamic by making specific future states easier to predict. We make use of this to propose a fully model-based algorithm that learns a discriminative world model, plans to explore efficiently with only a count-based intrinsic reward, and can subsequently plan to reach any discovered (abstract) states.

View on arXiv
Comments on this paper