Dario Amodei on Building the Most Trusted AI Lab
The Anthropic CEO discusses Constitutional AI, the race for AGI, and why safety isn't the enemy of capability.
Dario Amodei doesn’t look like someone who believes he’s building one of the most powerful technologies in human history. Sitting in Anthropic’s San Francisco office — modest by tech standards, all whiteboards and standing desks — he’s calm, precise, and disarmingly honest about what he doesn’t know.
“I think there’s maybe a 10-15% chance that AI goes really badly for humanity,” he says, unprompted. “And I think there’s maybe a 30-40% chance it goes extraordinarily well — better than almost anyone imagines. The work is about narrowing the bad outcomes and widening the good ones.”
This is vintage Amodei: probabilistic, unflinching, and motivated by a conviction that the best way to ensure AI safety is to be at the frontier building it.
The Anthropic Philosophy
Anthropic was founded in 2021 by Amodei and his sister Daniela, along with several other ex-OpenAI researchers, on a bet that safety and capability weren’t opposing forces. Three years later, that bet appears to be paying off.
Claude, Anthropic’s AI assistant, has become the model of choice for enterprises that care about reliability, accuracy, and behavioral predictability. Constitutional AI — the company’s approach to training models that follow principles rather than just human feedback — has proven remarkably effective at reducing harmful outputs without sacrificing capability.
“The insight was simple,” Amodei explains. “If you train a model to follow principles — be honest, be helpful, avoid harm — it turns out those principles make the model better at its job, not worse. Honesty and capability are correlated, not opposed.”
On the Race to AGI
Amodei is careful with the term “AGI” but doesn’t shy away from the trajectory.
“We’re building systems that are increasingly general. Claude can reason, plan, write code, analyze data, engage in nuanced conversation. Each generation is substantially more capable. Whether you call the destination ‘AGI’ or something else, the direction is clear.”
Does he think Anthropic will get there first?
“I think the question of ‘who gets there first’ is less important than ‘who gets there safely.’ If someone else builds a safe, beneficial AGI before us, that’s a good outcome. What I lose sleep over is someone building an unsafe one.”
The Business of Safety
Anthropic has raised over $15 billion and is valued at $60 billion — numbers that would have been unthinkable for a company whose founding thesis was “AI safety.” But Amodei insists that safety is the business strategy, not a constraint on it.
“Enterprise customers don’t want a model that’s occasionally brilliant and occasionally unhinged. They want reliability. They want predictability. They want to trust the system. Our safety research directly produces those qualities. Safety IS the product.”
The evidence supports this. Anthropic’s enterprise revenue has grown from essentially zero in early 2024 to a reported $4 billion annual run rate. Major customers include Amazon, which invested $4 billion in the company, and a growing list of Fortune 500 firms.
Looking Forward
What does Amodei think the next five years hold?
“AI systems will become dramatically more capable. They’ll be able to conduct genuine scientific research, manage complex projects, and engage in creative work that’s indistinguishable from human output. The economic impact will be enormous — think trillions of dollars of productivity gains.”
“But the risks scale with the capabilities. More powerful systems can cause more harm if misaligned. That’s why the safety work has to lead, not follow. You don’t add brakes to a car after you build the engine. You design them together.”
He pauses. “I’m fundamentally optimistic. The technology we’re building has the potential to solve problems that have plagued humanity for centuries — disease, poverty, climate change. But that potential is only realized if we build it right.”
Building it right. That’s the Anthropic bet. And three years in, it’s looking like a good one.