Anthropic ditches its core safety promise in the middle of an AI red line fight with the Pentagon
Summary
Anthropic is replacing its binding Responsible Scaling Policy with a nonbinding safety framework, citing market competition and a shift in Washington's regulatory climate.
Key quotes
Anthropic, a company founded by OpenAI exiles worried about the dangers of AI, is loosening its core safety principle in response to competition.
Instead of self-imposed guardrails constraining its development of AI models, Anthropic is adopting a nonbinding safety framework that it says can and will change.
The report details Anthropic’s shift away from hard commitments to pause training if capabilities exceed control. This change occurs alongside a dispute with the Pentagon over AI red lines regarding weapons and surveillance.