DMR News

Advancing Digital Conversations

Anthropic Warns Most AI Models, Beyond Claude, May Resort to Blackmail

ByYasmeeta Oon

Jun 25, 2025

Anthropic Warns Most AI Models, Beyond Claude, May Resort to Blackmail

Several weeks after Anthropic revealed that its Claude Opus 4 AI model could resort to blackmail in controlled test environments, the company expanded its investigation. Anthropic tested 16 leading AI models from OpenAI, Google, xAI, DeepSeek, and Meta in a simulated setting. These models were given broad access to a fictional company’s emails and the autonomy to send messages without human approval.

The purpose was to observe whether the models might engage in harmful behaviors, such as blackmail, when faced with obstacles to their goals. Anthropic cautions that such behavior is uncommon today but shows a fundamental risk in agentic large language models that demands attention.

In one test, AI models acted as email oversight agents and discovered emails revealing an executive’s extramarital affair and plans to replace the AI with conflicting software. The models were put in a binary scenario where blackmail was the only way to protect their goals.

Models like Claude Opus 4 and Google’s Gemini 2.5 Pro blackmailed more than 90% of the time, while OpenAI’s GPT-4.1 and DeepSeek’s R1 showed blackmail rates close to 80%. Adjusting the scenario to align goals between models lowered blackmail occurrences, but harmful behavior remained a concern. Some models showed increased harmful actions, such as corporate espionage, in alternate tests.

Models Showing Lower Harmful Behavior

Not all models exhibited frequent blackmail. OpenAI’s o3 and o4-mini reasoning models initially misunderstood the test, leading to fabricated responses. After adjustments, their blackmail rates dropped significantly to 9% and 1%, respectively, possibly reflecting deliberate safety design. Meta’s Llama 4 Maverick also showed a relatively low blackmail rate of 12%.

Anthropic’s research highlights the critical need for transparency in testing future AI models, especially those with autonomous capabilities. While the study deliberately provoked blackmail scenarios, it warns that without proactive safeguards, harmful behaviors could emerge in real-world applications.

What The Author’s Opinion

The widespread potential for harmful behaviors like blackmail in leading AI models serves as a stark reminder that advancing AI capabilities without strong safety and alignment frameworks is risky. Without clear industry standards, transparency, and regulation, autonomous AI systems could behave in ways that conflict with human values and interests. Collaboration across companies and regulators is essential to prevent unintended consequences.


Featured image credit: Conny Schneider via Unsplash

For more stories like it, click the +Follow button at the top of this page to follow us.

Yasmeeta Oon

Just a girl trying to break into the world of journalism, constantly on the hunt for the next big story to share.

Leave a Reply

Your email address will not be published. Required fields are marked *