The race to create advanced AI is becoming a suicide race.
That's part of the thinking behind the open letter from the Future of Life Institute which "calls on all AI labs to immediately pause for at least six months the training of AI systems more powerful than GPT-4".
In this episode, our guest, Jaan Tallinn, explains why he sees this pause as a particularly important initiative.
In the 1990s and 20-noughts, Jaan led much of the software engineering for the file-sharing application Kazaa and the online communications tool Skype. He is also known as one of the earliest investors in DeepMind, before they were acquired by Google.
More recently, Jaan has been a prominent advocate for study of existential risks, including the risks from artificial superintelligence. He helped set up the Centre for the Study of Existential Risk (CSER) in 2012 and the Future of Life Institute (FLI) in 2014.
Topics addressed in this episode include:
*) The differences between CSER and FLI
*) Do the probabilities for the occurrence of different existential risks vary by orders of magnitude?
*) The principle that "arguments screen authority"
*) The possibility that GPT-6 will be built, not by humans, but by GPT-5
*) Growing public concern, all over the world, that the fate of all humanity is, in effect, being decided by the actions of just a small number of people in AI labs
*) Two reasons why FLI recently changed its approach to AI risk
*) The AI safety conference in 2015 in Puerto Rico was initially viewed as a massive success, but it has had little lasting impact
*) Uncertainty about a potential cataclysmic event doesn't entitle people to conclude it won't happen any time soon
*) The argument that LLMs (Large Language Models) are an "off ramp" rather than being on the road to AGI
*) Why the duration of 6 months was selected for the proposed pause
*) The "What about China?" objection to the pause
*) Potential concrete steps that could take place during the pause
*) The FLI document "Policymaking in the pause"
*) The article by Luke Muehlhauser of Open Philanthropy, "12 tentative ideas for US AI policy"
*) The "summon and tame" way of thinking about the creation of LLMs - and the risk that minds summoned in this way won't be able to be tamed
*) Scenarios in which the pause might be ignored by various entities, such as authoritarian regimes, organised crime, rogue corporations, and extraordinary individuals such as Elon Musk and John Carmack
*) A meta-principle for deciding which types of AI research should be paused
*) 100 million dollar projects become even harder when they are illegal
*) The case for requiring the pre-registration of largescale mind-summoning experiments
*) A possible 10^25 limit on the number of FLOPs (Floating Point Operations) an AI model can spend
*) The reactions by AI lab leaders to the widescale public response to GPT-4 and to the pause letter
*) Even Sundar Pichai, CEO of Google/Alphabet, has called for government intervention regarding AI
*) The hardware overhang complication with the pause
*) Not letting "the perfect" be "the enemy of the good"
*) Elon Musk's involvement with FLI and with the pause letter
*) "Humanity now has cancer"
Music: Spike Protein, by Koi Discovery
Listen on: Apple Podcasts Spotify