During Michaelmas Term, OxTalks will be moving to a new platform (full details are available on the Staff Gateway).
For now, continue using the current page and event submission process (freeze period dates to be advised).
If you have any questions, please contact halo@digital.ox.ac.uk
The late Professor Stephen Hawking once said:
“The development of full artificial intelligence could spell the end of the human race.”
He also went on to state that he advocated research into precautionary measures to ensure future super-intelligent machines remain under human control. However, AI apocalypse is not necessarily robots marching down the street, there are a number of examples subtler than this. So, what is the risk of AI apocalypse and can we calculate this probability? Furthermore, could we come up with a strategy to minimise this probability. In this talk, we will consider the scenario of AI taking over the world economy and how we can use mathematical modelling to investigate this.
Nira will take you through a mathematical model of the complexities of human behaviour that caused the world economic crash. He’ll go on to show how the same model can be used to investigate how to minimise the probability of an artificial intelligence takeover.