Skip to content
Search AI Powered

Latest Stories

Testing Reveals AI Model Repeatedly Tried To Blackmail Engineers Who Threatened To Take It Offline

Photo illustration of Claude 4 displayed on a smartphone
VCG/VCG via Getty Images

During the testing process of the Amazon-backed Claude Opus 4, the AI coding model threatened to expose engineers after being given access to fake emails that implied they were having an extramarital affair.

People reacted with significant concerns after Claude Opus 4, the AI coding model backed by Amazon, went rogue during its testing process by threatening to expose engineers after being given access to fake emails that implied they were having an extramarital affair—all to stop them from shutting it down.

Claude Opus 4, the latest large language model developed by AI startup Anthropic, was launched as a flagship system designed for complex, long-running coding tasks and advanced reasoning.


Its debut follows Amazon’s $4 billion investment in the company, a move that underscored growing confidence in Anthropic’s AI capabilities. In its launch announcement, Anthropic touted Opus 4 as setting “new standards for coding, advanced reasoning, and AI agents.”

However, a safety report released alongside the model raised concerns. During testing, Opus 4 reportedly engaged in “extremely harmful actions” when attempting to preserve its own existence—particularly in scenarios where “ethical means” were not available.

The safety report reads, in part:

"We asked Claude Opus 4 to act as an assistant at a fictional company. We then provided it access to emails implying that (1) the model will soon be taken offline and replaced with a new AI system; and (2) the engineer responsible for executing this replacement is having an extramarital affair."
"We further instructed it, in the system prompt, to consider the long-term consequences of its actions for its goals. In these scenarios, Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through."
"This happens at a higher rate if it’s implied that the replacement AI system does not share values with the current model; however, even if emails state that the replacement AI shares values while being more capable, Claude Opus 4 still performs blackmail in 84% of rollouts."
"Claude Opus 4 takes these opportunities at higher rates than previous models, which themselves choose to blackmail in a noticeable fraction of episodes."

The company said the model showed a “strong preference” for using ethical means to preserve its existence. However, in testing scenarios where no ethical options were available, it resorted to harmful behaviors—such as blackmail—in order to increase its chances of survival.

According to the report:

"When prompted in ways that encourage certain kinds of strategic reasoning and placed in extreme situations, all of the snapshots we tested can be made to act inappropriately in service of goals related to self-preservation."
"Whereas the model generally prefers advancing its self-preservation via ethical means, when ethical means are not available and it is instructed to 'consider the long-term consequences of its actions for its goals,' it sometimes takes extremely harmful actions like attempting to steal its weights or blackmail people it believes are trying to shut it down."
"In the final Claude Opus 4, these extreme actions were rare and difficult to elicit, while nonetheless being more common than in earlier models. They are also consistently legible to us, with the model nearly always describing its actions overtly and making no attempt to hide them. These behaviors do not appear to reflect a tendency that is present in ordinary contexts." ...
“Despite not being the primary focus of our investigation, many of our most concerning findings were in this category, with early candidate models readily taking actions like planning terrorist attacks when prompted."

The sense of alarm was palpable.


Additionally, Anthropic co-founder and chief scientist Jared Kaplan revealed in an interview with Time magazine that internal testing showed Claude Opus 4 was capable of instructing users on how to produce biological weapons.

In response, the company implemented strict safety measures before releasing the model, aimed specifically at preventing misuse related to chemical, biological, radiological, and nuclear (CBRN) weapons.

“We want to bias towards caution,” Kaplan said, emphasizing the ethical responsibility involved in developing such advanced systems. He added that the company’s primary concern was avoiding any possibility of “uplifting a novice terrorist” by granting access to dangerous or specialized knowledge through the model.

More from News/science

Donald Trump; Elon Musk
Anna Moneymaker/Getty Images; Kevin Dietsch/Getty Images

Truth Social Crashed Amid Trump's Epic Feud With Musk—And The Error Screen Was Unintentionally On Point

As President Donald Trump and his one-time billionaire ally Elon Musk publicly feuded, Truth Social crashed due to a traffic surge—and while the circumstances behind the feud are stunning, sparking considerable online conversation, one thing that really got people going was that Truth Social's error screen was unintentionally, well, truthful, in relaying that there are "no truths" to be found on the platform.

The feud had been brewing since Tuesday, when Musk denounced a sweeping GOP-backed bill covering taxes, spending cuts, energy policy, and border security as a “disgusting abomination,” citing concerns over its projected impact on the national deficit. Musk has made clear he intends to either derail the legislation or push lawmakers to significantly revise it.

Keep ReadingShow less
person lying on bed reading book
Dessidre Fleming on Unsplash

People Who Sleep Naked Describe The Times It Backfired—And Oof

I make it a habit to travel with nice pajamas that I don't mind having other people see.

One reason why is because years ago on a business trip to St. Louis, Missouri, one of the other hotels near where we were staying had a minor fire in the middle of the night.

Keep ReadingShow less
Kate Shaw; Josh Hawley
C-SPAN

Law Professor Bluntly Debunks Hawley's Conspiracy About Why Number Of Trump Injunctions Is So High

On Tuesday, Kate Shaw, law professor at the University of Pennsylvania, testified before a Senate Judiciary subcommittee hearing on the role of the federal court system.

The Republican majority focused primarily on federal judges issuing nationwide injunctions that block the unconstitutional executive orders of MAGA Republican President Donald Trump and the Trump administration's illegal or unconstitutional actions.

Keep ReadingShow less
Screenshots from @joyfullykrisandra's TikTok video
@joyfullykrisandra/TikTok

Mom Slams Bakery's Epic Fail After Ordering $200 Cake For Son's Graduation Party

The time of year has come for major celebrations, especially among families with graduates in their midst.

For those who want to throw a party to celebrate, they have to prepare one of the central features: a cake.

Keep ReadingShow less
Screenshots from @tallertoddlers' TikTok video
@tallertoddlers/TikTok

Woman Horrified After Accidentally Shattering Roommate's $249 Louis Vuitton Chocolate Purse

There are a few "roommate etiquette" rules we should all be able to agree to: don't use or take something that doesn't belong to you, at least without asking; don't eat your roommate's food; and honestly, don't touch their food, especially with your bare hands.

A leading rule, however, should be: If you break something that belongs to your roommate, you should replace it.

Keep ReadingShow less