25% Chance That Things Could Go Really Badly With AI: Anthropic CEO Dario Amodei

AI is providing many remarkable breakthroughs, like gold medals in Math Olympiads and top coding competitions, but there’s also a chance that things could go extremely wrong.

Anthropic CEO Dario Amodei has offered a stark assessment of AI’s future prospects. The former OpenAI research executive revealed his personal probability estimates for how AI development might unfold—and his numbers are surprisingly stark for someone leading one of the world’s premier AI safety companies.

“I definitely think between the autonomous danger of the model and ending up on the bad side of some national security trade-offs and a job thing that goes in a very bad direction,” Amodei said in an interview with Axios, outlining the key risk categories that keep him awake at night. The Anthropic CEO, who has spent years thinking deeply about AI alignment and safety, wasn’t mincing words about the potential downsides.

His assessment gets even more striking when he puts numbers to these concerns: “I’m relatively an optimist, so I think there’s a 25% chance that things go really, really badly and a 75% chance that things go really, really well—with not much space” in between.

This binary outlook—where AI either transforms society for the better or creates catastrophic problems—reflects a growing consensus among AI researchers that we’re approaching a critical juncture. Amodei’s three-pronged risk assessment encompasses autonomous AI systems that could operate beyond human control, geopolitical tensions that could militarize AI development in dangerous ways, and economic disruption severe enough to destabilize society.

The implications of Amodei’s assessment are particularly significant given his position at the forefront of AI safety research. Anthropic, founded in 2021 by former OpenAI executives including Amodei and his sister Daria, has positioned itself as a safety-first AI company, developing Constitutional AI techniques and focusing heavily on alignment research. When someone dedicating their career to making AI safer still sees a 25% chance of catastrophic outcomes, it underscores the magnitude of the challenges ahead.

Recent developments lend credence to Amodei’s concerns. The rapid advancement of AI capabilities has outpaced regulatory frameworks, with models like GPT-4 and Claude demonstrating emergent abilities that researchers didn’t fully predict. Meanwhile, the intensifying AI race between the United States and China has raised fears about corners being cut in the rush to achieve artificial general intelligence first. On the economic front, Goldman Sachs estimates that AI could displace up to 300 million jobs globally, potentially creating social upheaval if the transition isn’t managed carefully.

Elon Musk too has a similar outlook on how AI could impact humanity. He’s previously said that there’s a 80% chance that it could lead to extreme prosperity, and a 10-20% chance that it could end humanity. These numbers seem to broadly line up with what Amodei seems to be predicting. What makes these predictions sobering are their stark splits between two very different kinds of outcomes with little middle ground. This suggests that AI leaders seem to believe we’re not just facing gradual challenges to be managed—we’re approaching a pivotal moment that will largely determine whether AI becomes humanity’s greatest achievement or its greatest mistake.

Posted in AI