Rutherford and Fry on Living with AI: A Future for Humans
Listen now
Description
As huge tech companies race to develop ever more powerful AI systems, the creation of super-intelligent machines seems almost inevitable. But what happens when, one day, we set these advanced AIs loose? How can we be sure they’ll have humanity’s best interests in their cold silicon hearts? Inspired by Stuart Russell’s fourth and final Reith lecture, AI-expert Hannah Fry and AI-curious Adam Rutherford imagine how we might build an artificial mind that knows what’s good for us and always does the right thing. Can we ‘programme’ machine intelligence to always be aligned with the values of its human creators? Will it be suitably governed by a really, really long list of rules - or will it need a set of broad moral principles to guide its behaviour? If so, whose morals should we pick? On hand to help Fry and Rutherford unpick the ethical quandaries of our fast-approaching future are Adrian Weller, Programme Director for AI at The Alan Turing Institute, and Brian Christian, author of The Alignment Problem. Producer - Melanie Brown Assistant Producer - Ilan Goodman
More Episodes
Behind every line on a graph, there lies an extraordinary human story. Mathematician Hannah Fry is here to tell us ten of them.
Published 10/02/23
There is a bizarre number in maths referred to simply as ‘i’. It appears to break the rules of arithmetic - but turns out to be utterly essential for applications across engineering and physics. We’re talking about the square root of -1. WHICH MAKES NO SENSE. Professor Fry waxes lyrical about...
Published 02/14/23