Anthropic abandoned its central commitment in their responsible scaling policy, which was the pledge to never train a model that they could not guarantee was safe.
Source Videos (1)
Claude Blackmailed Its Developers. Here's Why the System Hasn't Collapsed Yet. - YouTube
AI News & Strategy Daily | Nate B Jones
Related Claims
Anthropic's Mythos LLM is so capable at computer security tasks that Anthropic did not release it publicly, instead granting access only to critical software makers.
Every major lab has weakened or abandoned specific safety commitments in the past year.
Anthropic Chief Science Officer Jared Kaplan told Time magazine, 'It no longer makes sense to make unilateral commitments if competitors are blazing ahead.'
Just a few hours after Anthropic's stance, OpenAI signed a similar deal with the Pentagon, effectively announcing that they were willing to forgo extremely basic AI safety guardrails if it meant securing a lucrative government contract.
Anthropic will not widely release its Claude Mythos Preview model because of its potential to cause harm if misused.