Our research
How should we navigate explosive AI progress?
Featured
Preparing for the Intelligence Explosion
William MacAskill & Fin Moorhouse
March 2025
AI that can accelerate research could drive a century of technological progress over just a few years. During such a period, new technological or political developments will raise consequential and hard-to-reverse decisions, in rapid succession. We call these developments grand challenges.
These challenges include new weapons of mass destruction, AI-enabled autocracies, races to grab offworld resources, and digital beings worthy of moral consideration, as well as opportunities to dramatically improve quality of life and collective decision-making.
We argue that these challenges cannot always be delegated to future AI systems, and suggest things we can do today to meaningfully improve our prospects. AGI preparedness is therefore not just about ensuring that advanced AI systems are aligned: we should be preparing, now, for the disorienting range of developments an intelligence explosion would bring.
AI-Enabled Coups: How a Small Group Could Use AI to Seize Power
Tom Davidson, Lukas Finnveden & Rose Hadshar
April 2025
The development of AI that is more broadly capable than humans will create a new and serious threat: AI-enabled coups. An AI-enabled coup could be staged by a very small group, or just a single person, and could occur even in established democracies. Sufficiently advanced AI will introduce three novel dynamics that significantly increase coup risk. Firstly, military and government leaders could fully replace human personnel with AI systems that are singularly loyal to them, eliminating the need to gain human supporters for a coup. Secondly, leaders of AI projects could deliberately build AI systems that are secretly loyal to them, for example fully autonomous military robots that pass security tests but later execute a coup when deployed in military settings. Thirdly, senior officials within AI projects or the government could gain exclusive access to superhuman capabilities in weapons development, strategic planning, persuasion, and cyber offense, and use these to increase their power until they can stage a coup. To address these risks, AI projects should design and enforce rules against AI misuse, audit systems for secret loyalties, and share frontier AI systems with multiple stakeholders. Governments should establish principles for government use of advanced AI, increase oversight of frontier AI projects, and procure AI for critical systems from multiple independent providers.
Better Futures
Series
William MacAskill
August 2025
Suppose we want the future to go better. What should we do?
One approach is to avoid near-term catastrophes, like human extinction. This essay series explores a different, complementary, approach: improving on futures where we survive, to achieve a truly great future.
Will AI R&D Automation Cause a Software Intelligence Explosion?
Daniel Eth & Tom Davidson
March 2025
AI companies are increasingly using AI systems to accelerate AI research and development. Today’s AI systems help researchers write code, analyze research papers, and generate training data. Future systems could be significantly more capable – potentially automating the entire AI development cycle from formulating research questions and designing experiments to implementing, testing, and refining new AI systems. We argue that such systems could trigger a runaway feedback loop in which they quickly develop more advanced AI, which itself speeds up the development of even more advanced AI, resulting in extremely fast AI progress, even without the need for additional computer chips. Empirical evidence on the rate at which AI research efforts improve AI algorithms suggests that this positive feedback loop could overcome diminishing returns to continued AI research efforts. We evaluate two additional bottlenecks to rapid progress: training AI systems from scratch takes months, and improving AI algorithms often requires computationally expensive experiments. However, we find that there are possible workarounds that could enable a runaway feedback loop nonetheless.
AI Tools for Existential Security
Lizka Vaintrob & Owen Cotton-Barratt
March 2025
Humanity is not prepared for the AI-driven challenges we face. But the right AI tools could help us to anticipate and work together to meet these challenges — if they’re available in time. We can and should accelerate these tools.
Key applications include (1) epistemic tools, which improve human judgement; (2) coordination tools, which help diverse groups work identify and work towards shared goals; (3) risk-targeted tools to address specific challenges.
We can accelerate important tools by investing in task-relevant data, lowering adoption barriers, and securing compute for key R&D. While background AI progress limits potential gains, even small speedups could be decisive.
This is a priority area. There is lots to do already, and there will quickly be more. We should get started, and we should plan for a world with abundant cognition.
Stay up to date with our research
Stay up to speed with the latest research on preparing for transformative AI. Roughly weekly.
By subscribing you agree to Substack's terms of service. Unsubscribe anytime. Archives.
Research
Filter
(64)
AI for AI for Epistemics
Owen Cotton-Barratt & Lukas Finnveden

Abstract
AI could help us to build stronger AI-powered systems to help people track what is true. This brings important opportunities and risks.
Authors
Owen Cotton-Barratt & Lukas Finnveden
Topic
Differential AI acceleration
AI should (sometimes) be proactively prosocial
Tom Davidson & William MacAskill

Abstract
Should AI proactively take prosocial actions? Some argue we should aim for AI that is primarily “steerable”: simply enacting the will of its user. We disagree. As AI systems grow more autonomous and become embedded in economic and political life, their cumulative behavioral tendencies will shape society’s trajectory. AI that notices and acts on opportunities to benefit society more broadly could matter enormously. Despite some reasonable objections, we think the case for instilling proactive prosocial drives in AI is compelling.
Authors
Tom Davidson & William MacAskill
Topic
Macrostrategy
Concrete Projects in AGI Preparedness
William MacAskill & Fin Moorhouse

Abstract
We think there are lots of promising, neglected, and concrete projects that could help make the transition to superintelligence go better. This article describes some project ideas that readers might not have considered much before. They include: AI character evaluations, automated macrostrategy, security audits for AI sabotage, tools for improving collective epistemics and enabling coordination, and more.
Authors
William MacAskill & Fin Moorhouse
Topic
Macrostrategy
The importance of AI character
William MacAskill & Tom Davidson

Abstract
We expect AI “character” (e.g. how obedient, honest, cooperative, or altruistic AIs are, and in what circumstances) to have a big effect on society and on how well the future goes. We think that figuring out what characters AI systems should have, and getting companies to actually build them that way, is among the most valuable things that people can do today.
Authors
William MacAskill & Tom Davidson
Topic
Macrostrategy
Broad Timelines
Toby Ord

Abstract
No-one is sure when AI will begin having transformative impacts upon the world, and we shouldn’t be sure: there just isn’t enough evidence to pin it down. But we don’t need to wait for certainty in order to act and plan with epistemic humility. What does wise planning look like in a world of deeply uncertain AI timelines? I conclude that taking uncertainty seriously has real implications for how both individuals and groups can contribute to making this AI transition go well.
Author
Toby Ord
Topic
Modelling AI progress
Should We Lock in Post-AGI Agreements Under Uncertainty?
Fin Moorhouse

Abstract
It’s a widely-held view that we should delay locking-in consequential or long-lasting agreements until we know much more about the world after a possible intelligence explosion, and until we can engage in extensive reflection on our beliefs and values. But some mutually-beneficial deals depend on shared uncertainty about the future. If we prevent those deals, we’re potentially leaving a lot of expected value off the table. So we need to make those deals before we have time to seriously reflect. I discuss when this could be true, and what the practical implications could be.
Author
Fin Moorhouse
Topic
Macrostrategy
Moral public goods are a big deal for whether we get a good future
Tom Davidson, William MacAskill & Mia Taylor

Abstract
A moral public good is something many people want to exist for moral reasons (for example, poverty reduction). Even if a group of people each care more about idiosyncratic selfish goods than they do moral public goods, they could still all be better off if they coordinate to fund moral public goods. Ensuring that this kind of coordination happens between future people, who might control large amounts of the universe's available resources, could be very important for determining how well the future goes.
Authors
Tom Davidson, William MacAskill & Mia Taylor
Topic
Macrostrategy
Design sketches: tools for strategic awareness
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Part 3 of Design sketches for a more sensible world

Abstract
Near-term AI could be used to power technologies that give individuals and organizations a deeper strategic awareness of the world around them, helping them spot opportunities and avoid pitfalls as they make plans. We think improved strategic awareness could be especially important for empowering humanity to handle the challenges that advanced AI is likely to bring. Here we sketch three technologies that build towards this vision.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Macrostrategy & Differential AI acceleration
Design sketches for a more sensible world
Series
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar

Abstract
We think that near-term AI systems could transform our ability to reason and coordinate, significantly improving our chances of safely navigating the transition to advanced AI systems. This sequence gives a series of design sketches for specific technologies that we think could help. We hope that these sketches make a more sensible world easier to envision, and inspire people to start building the relevant tech.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Modelling AI progress, Differential AI acceleration & Macrostrategy
Design Sketches: Angels-on-the-Shoulder
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Part 2 of Design sketches for a more sensible world

Abstract
Near-term AI could allow us to build many technological analogues to ‘angels-on-the-shoulder’: highly customized tools that help people to better navigate their environments or handle tricky situations in ways they’ll feel good about later. These could mean more endorsed decisions, and fewer unforced errors. Here we sketch five technologies that build towards this vision.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Macrostrategy & Differential AI acceleration
The International AGI Project Series
Series
William MacAskill

Abstract
This is a series of papers and research notes on the idea that AGI should be developed as part of an international collaboration between governments. We aim to (i) assess how desirable an international AGI project is; (ii) assess what the best version of an international AGI project (taking feasibility into account) would look like.
Author
William MacAskill
Topic
International governance
What an international project to develop AGI should look like
William MacAskill
Part 1 of The international AGI project series

Abstract
What would the best version of an international project to develop AGI look like? In this research note, I set out my tentative best guess: “Intelsat for AGI”. This would be a US-led international project modelled on Intelsat (an international project that set up the first global communications satellite network), with broad benefit sharing for non-members. The primary case is that, within the domain of international AGI projects, this looks unusually feasible, and yet it would significantly reduce catastrophic risk compared to a US-only project.
Author
William MacAskill
Topic
International governance
Show all