Anthropic Drops Flagship Safety Pledge

Slashdot
by msmash
February 25, 2026
AI-Generated Deep Dive Summary
Anthropic, known for its strong focus on AI safety, has abandoned a key part of its 2023 Responsible Scaling Policy, which previously required the company to ensure safety measures were adequate before training any AI system. The decision reflects a shift in strategy as the company adapts to the rapid pace of AI advancements and competition. Instead of making unilateral commitments, Anthropic now plans to match or exceed competitors' safety efforts while only delaying development if it believes catastrophic risks are significant and it is leading the AI race. This new approach signals a more agile and competitive stance on AI safety. The revised policy was approved by CEO Dario Amodei and the board, with Chief Science Officer Jared Kaplan explaining that Anthropic felt it could no longer sustain its previous strict position as competitors advanced faster. The company now prioritizes dynamic risk assessment and mitigation, acknowledging that existing methods are struggling to keep up with AI's rapid evolution. To maintain transparency, Anthropic will publish "Risk Reports" every three to six months and release "Frontier Safety Roadmaps" outlining future safety goals. Chris Painter, director of policy at the AI evaluation nonprofit METR, reviewed an early draft of the updated policy and noted that it reflects a recognition of the need for triage mode in safety planning. He emphasized that the shift underscores how quickly the field is evolving and the challenges in keeping pace with emerging risks. This new strategy positions Anthropic to stay competitive while addressing potential dangers posed by advancing AI technologies. For readers interested in tech, this development highlights the growing emphasis on balancing innovation with caution in AI research. As the industry races forward, companies like Anthropic are reevaluating their approaches to safety to ensure they remain relevant and effective. The move also raises questions about how other players in the field will respond, potentially sparking a broader conversation about responsible AI development and its future trajectory.
Verticals
tech
Originally published on Slashdot on 2/25/2026