OxTalks will soon move to the new Halo platform and will become 'Oxford Events.' There will be a need for an OxTalks freeze. This was previously planned for Friday 14th November – a new date will be shared as soon as it is available (full details will be available on the Staff Gateway).
In the meantime, the OxTalks site will remain active and events will continue to be published.
If staff have any questions about the Oxford Events launch, please contact halo@digital.ox.ac.uk
In this talk we offer a generative perspective on value function approximation in reinforcement learning. Based on this perspective we develop the Compress and Control algorithm, which transforms arbitrary density estimators into value functions. In particular, we consider compression methods such as the Lempel-Ziv and Context Tree Switching algorithms as base models. The appeal of compression methods for density estimation is that they are in a sense feature-free: they can be tractably applied to bit sequences, and therefore to any kind of data. Along with a theoretical overview of the method, we present empirical results on the Atari 2600 platform.
Reference: webdocs.cs.ualberta.ca/~mg17/publications/veness14compress.pdf
—
Marc G. Bellemare received his Ph.D. from the University of Alberta, where he investigated the concept of domain-independent agents and led the design of the Arcade Learning Environment. His research interests include reinforcement learning, online learning, information theory, lifelong learning, and randomized algorithms. He is currently at Google DeepMind.
Joel Veness is a Senior Research Scientist at Google DeepMind. He is interested in reinforcement learning, universal source coding, Bayesian nonparametrics and game AI.