OxTalks will soon move to the new Halo platform and will become 'Oxford Events.' There will be a need for an OxTalks freeze. This was previously planned for Friday 14th November – a new date will be shared as soon as it is available (full details will be available on the Staff Gateway).
In the meantime, the OxTalks site will remain active and events will continue to be published.
If staff have any questions about the Oxford Events launch, please contact halo@digital.ox.ac.uk
Meta-learning, the ability to learn the right structure or prior knowledge to facilitate new learning, is heavily reliant on structured data. Humans, deep RL agents, and even large language models (LLMs) are all capable of meta-learning. While recurrent neural network-based models can be linked to neural activations in biological organisms, understanding how LLMs perform this quick, in-context learning is more difficult. LLMs can be pre-trained on human-generated artifacts, such as the internet and books, which contain substantial structure and enable good generalization. However, the lack of specific knowledge of the training data makes it challenging to quantify their performance, especially as they’re being increasingly deployed at scale in the real world. New approaches have been introduced that allow us to more closely interrogate how they work, approaches directly taken from the cognitive sciences. In this talk I discuss how we can better understand both deep RL agents and LLMs by looking at structure within their training data through this lens, and why they’re so powerful.