
AI Safety Fundamentals
Listen to resources from the AI Safety Fundamentals courses!
https://aisafetyfundamentals.com/
Episodes
172 episodes
d/acc: One Year Later
By Vitalik ButerinEthereum founder Vitalik Buterin describes how democratic, defensive and decentralised technologies could distribute AI's power across society rather than concentrating it, offering a middle path between unchecked techn...
•
43:12

A Playbook for Securing AI Model Weights
By Sella Nevo et al.In this report, RAND researchers identify real-world attack methods that malicious actors could use to steal AI model weights. They propose a five-level security framework that AI companies could implement to defend a...
•
19:56

AI Emergency Preparedness: Examining the Federal Government's Ability to Detect and Respond to AI-Related National Security Threats
By Akash Wasil et al.This paper uses scenario planning to show how governments could prepare for AI emergencies. The authors examine three plausible disasters: losing control of AI, AI model theft, and bioweapon creation. They then expos...
•
9:44

Resilience and Adaptation to Advanced AI
By Jamie BernardiJamie Bernardi argues that we can't rely solely on model safeguards to ensure AI safety. Instead, he proposes "AI resilience": building society's capacity to detect misuse, defend against harmful AI applications, and red...
•
13:42

Introduction to AI Control
By Sarah Hastings-WoodhouseAI Control is a research agenda that aims to prevent misaligned AI systems from causing harm. It is different from AI alignment, which aims to ensure that systems act in the best interests of their users. Put s...
•
10:19

The Project: Situational Awareness
By Leopold AschenbrennerA former OpenAI researcher argues that private AI companies cannot safely develop superintelligence due to security vulnerabilities and competitive pressures that override safety. He argues that a government-led '...
•
32:04

Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?
By Yoshua Bengio et al.This paper argues that building generalist AI agents poses catastrophic risks, from misuse by bad actors to a potential loss of human control. As an alternative, the authors propose “Scientist AI,” a non-agentic sy...
•
21:01

The Intelligence Curse
By Luke Drago and Rudolf LaineThis section explores how the arrival of AGI could trigger an “intelligence curse,” where automation of all work removes incentives for states and companies to care about ordinary people. It frames the trill...
•
2:19:08

AI Is Reviving Fears Around Bioterrorism. What’s the Real Risk?
By Kyle HiebertThe global spread of large language models is heightening concerns that extremists could leverage AI to develop or deploy biological weapons. While some studies suggest chatbots only marginally improve bioterror capabiliti...
•
8:28

AI and the Evolution of Biological National Security Risks
By Bill Drexel and Caleb WithersThis report considers how rapid AI advancements could reshape biosecurity risks, from bioterrorism to engineered superviruses, and assesses which interventions are needed today. It situates these risks in ...
•
16:13

The Intelligence Curse (Sections 1-3)
By Luke Drago and Rudolf LaineThis piece explores key arguments from sections 3 and 4 of The Intelligence Curse, continuing the authors’ analysis of how increasing intelligence can create paradoxical disadvantages, tradeof...
•
44:21

The Most Important Time in History Is Now
By Tomas PueyoThis blog post traces AI's rapid leap from high school to PhD-level intelligence in just two years, examines whether physical bottlenecks like computing power can slow this acceleration, and argues that recent efficiency br...
•
38:31

Why Do People Disagree About When Powerful AI Will Arrive?
By Sarah Hastings-WoodhouseMost experts agree that AGI is possible. They also agree that it will have transformative consequences. There is less consensus about what these consequences will be. Some believe AGI will usher in an age of ra...
•
22:03

Governance of Superintelligence
By Sam Altman, Greg Brockman, Ilya SutskeverOpenAI's leadership outline how humanity might govern superintelligence, proposing international oversight with inspection powers similar to nuclear regulation. They argue the AI systems arrivi...
•
5:22

Scaling: The State of Play in AI
By Ethan MollickThis post explains the "scaling laws" that drive rapid AI progress: when you make AI models bigger and train them with more computing power, they get smarter at most tasks. The piece also introduces a second scaling law, ...
•
24:39

Measuring AI Ability to Complete Long Tasks
By Thomas Kwa et al.We propose measuring AI performance in terms of the length of tasks AI agents can complete. We show that this metric has been consistently exponentially increasing over the past 6 years, with a doubling time of around...
•
15:06

The AI Revolution: The Road to Superintelligence
By Tim UrbanTim Urban uses historical analogies to show why AI progress might accelerate much faster than we expect, and how AI systems could rapidly self-improve from human-level to superintelligent capabilities.Source:
•
48:47

"Long" Timelines to Advanced AI Have Gotten Crazy Short
By Helen TonerHelen Toner, former OpenAI board member, reveals how the AI timeline debate has compressed: even conservative experts who once dismissed advanced AI concerns now predict human-level systems within decades. Rapid AI progress...
•
9:41

In Search of a Dynamist Vision for Safe Superhuman AI
By Helen TonerThis essay describes AI safety policies that rely on centralised control (surveillance, fewer AI projects, licensing regimes) as "stasist" approaches that sacrifice innovation for stability. Toner argues we need "dynamist" ...
•
16:55

It’s Practically Impossible to Run a Big AI Company Ethically
By Sigal Samuel (Vox Future Perfect)Even "safety-first" AI companies like Anthropic face market pressure that can override ethical commitments. This article demonstrates the constraints facing AI companies, and why voluntary corporate go...
•
17:14

Seeking Stability in the Competition for AI Advantage
By Iskander Rehman, Karl P. Mueller, Michael J. MazarrThis RAND article describes some of the international dynamics driving the race to AGI between the US and China, and analyses whether nuclear deterrence logic applies to this race.
•
18:18

Solarpunk: A Vision for a Sustainable Future
By Joshua KrookWhat might sustainable human progress look like, beyond pure technological acceleration? This essay provides an alternative vision, based on communities living in greater harmony with each other and with nature, alongside ...
•
13:21

The Gentle Singularity
By Sam AltmanThis blog post offers a vivid, optimistic vision of rapid AI progress from the CEO of OpenAI. Altman suggests that the accelerating technological change will feel "impressive but manageable," and that there are serious chall...
•
10:20

Preparing for Launch
By Tim Fist, Tao Burga, and Tim HwangThe Institute for Progress lays out how the US Government could shape the development of AI towards human flourishing by accelerating beneficial AI applications and defences against societal harms.
•
38:01

AI-Enabled Coups: How a Small Group Could Use AI to Seize Power
By Tom Davidson, Lukas Finnveden and Rose Hadshar. The development of AI that is more broadly capable than humans will create a new and serious threat: AI-enabled coups. An AI-enabled coup could be staged by a very small gr...
•
2:09:31
