Skip to content

Small Extinction Risk Warrants Extreme Caution

by Benjamin Mann on July 20, 2025

Benjamin Mann approaches AI safety as an existential imperative, believing that creating superintelligence might be humanity's final invention—with consequences that could last forever.

Mann views AI safety not as a theoretical concern but as a concrete necessity requiring immediate action. "My personal feeling is that things are overwhelmingly likely to go well, but on the margin almost nobody is looking at the downside risk, and the downside risk is very large." This perspective drives his work at Anthropic, where safety is the top priority rather than one competing interest among many.

He frames the stakes through a powerful analogy: "If I told you there's a one percent chance that the next time you got in an airplane you would die, you'd probably think twice even though it's only one percent because it's just such a bad outcome. If we're talking about the whole future of humanity, it's just a dramatic future to be gambling with."

Mann estimates the probability of an extremely bad outcome from AI is somewhere between 0-10%, but believes the marginal impact of working on safety is enormous precisely because so few people (less than a thousand worldwide) are focused on it. This creates a responsibility to "make triple sure that it's going to go well" rather than assuming safety will emerge naturally.

His approach to alignment is deeply empirical, focusing on building systems that understand human values and intentions rather than just following instructions literally. This manifests in Anthropic's Constitutional AI, which uses natural language principles derived from sources like the UN Declaration of Human Rights to guide model behavior.

For leaders and practitioners, Mann's perspective suggests several practical implications:

  1. Safety work should be integrated into development rather than added afterward—the personality and character of Claude is a direct result of alignment research, not a separate feature.

  2. The most valuable innovations may come from solving safety challenges that competitors can't address, creating unique opportunities: "Through our safety research we have a big opportunity to do things that no other company can safely do."

  3. When building AI systems, focus on understanding what people want, not just what they say—avoiding the "monkey paw" scenario where literal interpretation leads to harmful outcomes.

  4. Transparency about failures and risks builds trust with stakeholders and policymakers rather than undermining it: "If you talk to policymakers, they really appreciate this kind of thing because they feel like we're giving them the straight talk."

Mann's philosophy ultimately centers on responsibility—acknowledging that while progress is inevitable and likely positive, the stakes are too high to leave safety as an afterthought.