Our research

How should we navigate explosive AI progress?

Stay up to date with our research

Enter your email address to subscribe to our newsletter about transformative AI research
Stay up to speed with the latest research on preparing for transformative AI. Roughly weekly.
By subscribing you agree to Substack's terms of service. Unsubscribe anytime. Archives.

Listen

Listen on your favorite platform and never miss an episode.
Spotify
Apple Podcasts
YouTube
Podcast Addict
Pocket Casts
Overcast
RSS
Amazon Music
Pinecast

Research

Filter
(64)

AI for AI for Epistemics

Owen Cotton-Barratt & Lukas Finnveden
Abstract
AI could help us to build stronger AI-powered systems to help people track what is true. This brings important opportunities and risks.
Authors
Owen Cotton-Barratt & Lukas Finnveden
Topic
Differential AI acceleration

AI should (sometimes) be proactively prosocial

Tom Davidson & William MacAskill
Abstract
Should AI proactively take prosocial actions? Some argue we should aim for AI that is primarily “steerable”: simply enacting the will of its user. We disagree. As AI systems grow more autonomous and become embedded in economic and political life, their cumulative behavioral tendencies will shape society’s trajectory. AI that notices and acts on opportunities to benefit society more broadly could matter enormously. Despite some reasonable objections, we think the case for instilling proactive prosocial drives in AI is compelling.
Authors
Tom Davidson & William MacAskill
Topic
Macrostrategy

Concrete Projects in AGI Preparedness

William MacAskill & Fin Moorhouse
Abstract
We think there are lots of promising, neglected, and concrete projects that could help make the transition to superintelligence go better. This article describes some project ideas that readers might not have considered much before. They include: AI character evaluations, automated macrostrategy, security audits for AI sabotage, tools for improving collective epistemics and enabling coordination, and more.
Authors
William MacAskill & Fin Moorhouse
Topic
Macrostrategy

The importance of AI character

William MacAskill & Tom Davidson
Abstract
We expect AI “character” (e.g. how obedient, honest, cooperative, or altruistic AIs are, and in what circumstances) to have a big effect on society and on how well the future goes. We think that figuring out what characters AI systems should have, and getting companies to actually build them that way, is among the most valuable things that people can do today.
Authors
William MacAskill & Tom Davidson
Topic
Macrostrategy

Broad Timelines

Toby Ord
Abstract
No-one is sure when AI will begin having transformative impacts upon the world, and we shouldn’t be sure: there just isn’t enough evidence to pin it down. But we don’t need to wait for certainty in order to act and plan with epistemic humility. What does wise planning look like in a world of deeply uncertain AI timelines? I conclude that taking uncertainty seriously has real implications for how both individuals and groups can contribute to making this AI transition go well.
Author
Toby Ord
Topic
Modelling AI progress

Should We Lock in Post-AGI Agreements Under Uncertainty?

Fin Moorhouse
Abstract
It’s a widely-held view that we should delay locking-in consequential or long-lasting agreements until we know much more about the world after a possible intelligence explosion, and until we can engage in extensive reflection on our beliefs and values. But some mutually-beneficial deals depend on shared uncertainty about the future. If we prevent those deals, we’re potentially leaving a lot of expected value off the table. So we need to make those deals before we have time to seriously reflect. I discuss when this could be true, and what the practical implications could be.
Author
Fin Moorhouse
Topic
Macrostrategy

Moral public goods are a big deal for whether we get a good future

Tom Davidson, William MacAskill & Mia Taylor
Abstract
A moral public good is something many people want to exist for moral reasons (for example, poverty reduction). Even if a group of people each care more about idiosyncratic selfish goods than they do moral public goods, they could still all be better off if they coordinate to fund moral public goods. Ensuring that this kind of coordination happens between future people, who might control large amounts of the universe's available resources, could be very important for determining how well the future goes.
Authors
Tom Davidson, William MacAskill & Mia Taylor
Topic
Macrostrategy

Design sketches: tools for strategic awareness

Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Part 3 of Design sketches for a more sensible world
Abstract
Near-term AI could be used to power technologies that give individuals and organizations a deeper strategic awareness of the world around them, helping them spot opportunities and avoid pitfalls as they make plans. We think improved strategic awareness could be especially important for empowering humanity to handle the challenges that advanced AI is likely to bring. Here we sketch three technologies that build towards this vision.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Macrostrategy & Differential AI acceleration

Design sketches for a more sensible world

Series
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Abstract
We think that near-term AI systems could transform our ability to reason and coordinate, significantly improving our chances of safely navigating the transition to advanced AI systems. This sequence gives a series of design sketches for specific technologies that we think could help. We hope that these sketches make a more sensible world easier to envision, and inspire people to start building the relevant tech.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Modelling AI progress, Differential AI acceleration & Macrostrategy

Design Sketches: Angels-on-the-Shoulder

Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Part 2 of Design sketches for a more sensible world
Abstract
Near-term AI could allow us to build many technological analogues to ‘angels-on-the-shoulder’: highly customized tools that help people to better navigate their environments or handle tricky situations in ways they’ll feel good about later. These could mean more endorsed decisions, and fewer unforced errors. Here we sketch five technologies that build towards this vision.
Authors
Owen Cotton-Barratt, Lizka Vaintrob, Oly Sourbut & Rose Hadshar
Topics
Macrostrategy & Differential AI acceleration

The International AGI Project Series

Series
William MacAskill
Abstract
This is a series of papers and research notes on the idea that AGI should be developed as part of an international collaboration between governments. We aim to (i) assess how desirable an international AGI project is; (ii) assess what the best version of an international AGI project (taking feasibility into account) would look like.
Author
William MacAskill
Topic
International governance

What an international project to develop AGI should look like

William MacAskill
Part 1 of The international AGI project series
Abstract
What would the best version of an international project to develop AGI look like? In this research note, I set out my tentative best guess: “Intelsat for AGI”. This would be a US-led international project modelled on Intelsat (an international project that set up the first global communications satellite network), with broad benefit sharing for non-members. The primary case is that, within the domain of international AGI projects, this looks unusually feasible, and yet it would significantly reduce catastrophic risk compared to a US-only project.
Author
William MacAskill
Topic
International governance
Show all

Search

Search for articles...