Anthropic changes Safety guidelines, will now train AI model even if safety not guaranteed

Anthropic is currently fighting aggressively to win the AI race. With its model Claude the company is certainly at the forefront of the AI development. However, to keep its position and continue to remain competitive, it has now revised its AI safety policy and has dropped one of its most defining safety promises, the very core promise that led to its creation in 2021.

In its latest update to the Responsible Scaling Policy, which it had created in 2023, Anthropic has announced that it will no longer commit to stop training or releasing more powerful AI models solely on the basis of safety concerns.

Anthropic explains that while its earlier model helped it create better safeguards for current risks, it has hit a “zone of ambiguity”, where it is difficult to prove exactly when an AI model becomes dangerous. According to the company, “the science of model evaluation isn’t well-developed enough to provide dispositive answers,” thus it is difficult for Anthropic to convince other AI companies or the government to stop and wait for them.

The company also argues that if it sticks to strict “stop” rules while other companies continue building more powerful AI, it could end up falling behind. And if companies like Anthropic lose influence, that could make “a world that is less safe”. So instead of slowing itself down alone, Anthropic has decided to move away from rigid stop rules and adopt a more flexible approach.

Anthropic also noted that reaching the highest levels of security against major threats is currently “not possible” for a single company to achieve alone. So, instead of maintaining strict, unilateral commitments, it is shifting towards a more flexible system focused on transparency and regular “Risk Reports”. Through these reports, the company says it will clearly explain how it is identifying and managing risks as its AI models become more powerful.

Anthropic shakes the core safety rule?

Anthropic’s latest policy shift has come as a big surprise for the industry. Founded by former OpenAI researchers who were vocal about the risks of advanced AI, the company built its reputation on putting safety first. When it introduced its Responsible Scaling Policy (RSP) in 2023, the framework of the company centred on a clear “hard-stop” on the launch of models which do not guarantee safety. In fact that commitment became one of the key reasons why Anthropic is seen as the most safety-focused of the major AI labs.

But in middle of the heated competition in the AI race, the company has now decided to keep safety a somewhat second priority.

Latest

US judge drops Elon Musk fraud claims against OpenAI, trial to continue

A US judge dismissed Elon Musk's fraud claims against OpenAI and Sam Altman but allowed charitable trust and unjust enrichment claims to proceed to trial. The r

US orders global diplomatic push over alleged Chinese AI distillation

Washington has directed diplomats worldwide to raise concerns about Chinese firms, including DeepSeek, allegedly distilling US AI models. The cable sharpens the

AI smart glasses will help visually impaired runners take on the London Marathon

AI smart glasses will help visually impaired runners take on the London Marathon

Meta is adding parental controls to monitor AI use of teens on Facebook and Instagram

Meta has introduced a parental supervision tool, allowing parents to view the key topics their teens have discussed with Meta AI across Instagram, Messenger, an

What is Meta doing with employees? Random layoffs, micro-managing work, new appraisal system

Layoffs, AI tracking and a tougher review system mark Meta's latest reset

Topics

You’ve his number: Leavitt makes joke at Trump’s expense before her maternity leave

The remark was seen as a playful nod to Donald Trump’s habit of communicating directly and frequently, often bypassing traditional channels and keeping the me

Scott Bessent says US holds Hormuz blockade, rejects Iranian or Russian oil waivers

The United States will not renew sanctions waivers for certain Iranian and Russian oil cargoes. The decision tightens pressure on both producers as energy marke

US judge drops Elon Musk fraud claims against OpenAI, trial to continue

A US judge dismissed Elon Musk's fraud claims against OpenAI and Sam Altman but allowed charitable trust and unjust enrichment claims to proceed to trial. The r

US orders global diplomatic push over alleged Chinese AI distillation

Washington has directed diplomats worldwide to raise concerns about Chinese firms, including DeepSeek, allegedly distilling US AI models. The cable sharpens the

Struggling DC face stern test against unbeaten PBKS in clash of contrasts

IPL 2026, DC vs PBKS: Delhi Capitals need something close to their best to pull their campaign back on track when they take on a rampaging Punjab Kings in the I

Why should pacers have all the fun? Krunal Pandya on his wily variations in IPL 2026

Krunal Pandya has outlined how he has reshaped his bowling for RCB in the IPL. His new variations reflect how bowlers are adapting to keep pace with T20 batting

Selfless Virat Kohli praises Devdutt Padikkal as real hero of RCB’s win over GT

Virat Kohli credited Devdutt Padikkal after Royal Challengers Bengaluru chased down 206 against Gujarat Titans. Their partnership shaped the chase and kept RCB

UK mother, 56, dies at assisted dying clinic in Switzerland after son’s death

A 56-year-old woman from the UK has died at an assisted dying clinic in Switzerland, according to news report. She had earlier spoken about struggling with grie
spot_img

Related Articles

Popular Categories

spot_imgspot_img