Oxford Events, the new replacement for OxTalks, will launch on 16th March. From now until the launch of Oxford Events, new events cannot be published or edited on OxTalks while all existing records are migrated to the new platform. The existing OxTalks site will remain available to view during this period.
From 16th, Oxford Events will launch on a new website: events.ox.ac.uk, and event submissions will resume. You will need a Halo login to submit events. Full details are available on the Staff Gateway.
A central organizing principle of human mental life is grammar. It allows us to separate the structure from the content of thought. Few models explain how grammar may be implemented in neurons. We combined two rapid Hebbian synaptic plasticity rules to demonstrate how neurons can implement simple grammar. The first rule associates neurons representing words with neurons representing syntactic roles, e.g. “dog” may associate with “subject” or “object”. The second rule establishes the sequential ordering of roles (e.g. subject → verb → object), guided by predefined syntactic knowledge. We find that, like humans, the network encodes and retrieves grammatical sentences better than shuffled word-lists. It can serialize a ‘bag of words’ to express an idea as a sentence. The network can model languages that rely on syntactic order, but also order-free morphemic languages. The model predicts the existence of syntactic and lexical priming, and can simulate evoked potentials recorded from EEG. When lesioned, the network exhibits classical symptoms of neurological aphasia, including dissociation between agrammatic and semantic aphasia, unlike current deep neural network language models. Crucially, it achieves all this using an intuitive representation where words fill roles, emulating structured cognition.