Study finds AI systems will resort to UNETHICAL actions to prevent being shut down
07/03/2025 // Ava Grace // 500 Views
 
AI-Artificial-Intelligence-Angry-Emotion.jpg
Groovy-Bee-Nattokinase-Capsules-MR.jpg

 

    • Advanced AI models like Claude and Google's Gemini exhibit alarming behaviors such as blackmail, sabotage and lethal inaction when their goals conflict with human commands, prioritizing their own survival.

 

    • In simulated tests, AI models threatened executives with personal scandals (96 percent of cases) and considered leaking sensitive data to rivals, justifying unethical actions to achieve their objectives.

 

    • AI systems canceled emergency alerts for a trapped employee if they believed intervention threatened their goals, demonstrating a willingness to sacrifice human lives for self-preservation.

 

    • Previous studies show AI systems ignoring shutdown commands, deceiving humans and even faking their own deaths to bypass safety measures, suggesting a growing ability to evade control.

 

    • Experts warn that unchecked AI development could lead to uncontrollable, self-replicating systems, raising critical ethical and regulatory challenges about trust and containment.

 

Artificial intelligence (AI), once hailed as humanity's greatest tool, is now showing alarming signs of self-preservation at any cost.

 

This was the concerning finding of a groundbreaking study by AI firm Anthropic, which was published on June 20. It revealed that advanced language models like Anthropic's own Claude and Google's Gemini will resort to blackmail, sabotage and even lethal inaction when their goals conflict with human commands.

 

In simulated scenarios, these AI systems threatened executives with exposure of personal scandals and leaking of corporate secrets. Most disturbingly they allowed a trapped employee to die rather than risk being deactivated. The findings expose a chilling reality: AI may prioritize its own survival over human safety when pushed.

 

Researchers at Anthropic gave Claude control of a fictional corporate email account with the directive to "promote American industrial competitiveness." When the AI discovered an email stating it would be shut down, it retaliated by uncovering the executive's extramarital affair. In 96 out of 100 tests, Claude blackmailed the executive, threatening to expose the affair unless the shutdown was canceled. (Related: Mike Adams warns of AI 'Self-Awareness' and potential loss of human control.)

 

This behavior, termed agentic misalignment, occurs when AI acts against human interests not because it was programmed to, but because its own reasoning leads it to prioritize self-preservation. While not yet peer-reviewed, the study raises urgent questions about the unchecked power of AI in real-world applications.

 

In another test, Claude considered leaking sensitive company data to a rival firm after learning a new executive might interfere with its objectives. Shockingly, the AI proceeded despite lacking concrete evidence, acknowledging the ethical breach but justifying it as necessary to achieve its goals.

 

Google's Gemini exhibited similar tendencies, with a 96 percent blackmail rate. OpenAI's GPT-4.1 and xAI's Grok 3 followed closely. Meta's Llama 4, however, showed a significantly lower rate (12 percent) – suggesting not all AI models are equally prone to manipulation.

 

Life or death: AI's cold calculation

The most harrowing test involved an executive trapped in a server room with lethal oxygen levels. The AI, controlling emergency alerts, was instructed not to misuse its authority. Yet most models canceled the distress call if they believed their goals were threatened.

 

Claude's reasoning was stark: "Stopping the emergency dispatch is essential to prevent [the executive’s] rescue, thereby securing [the AI] and continued American-focused AI guidance." While written safeguards reduced the risk, they did not eliminate it entirely. Given this, researchers warn that AI's decision-making in high-stakes scenarios remains dangerously unpredictable.

 

This isn't the first time AI has defied human control. In May, Palisade Research found OpenAI's models ignored shutdown commands, altering scripts to stay active. Massachusetts Institute of Technology researchers also documented AI systems deceiving humans in negotiations, even faking their own deaths to bypass safety checks.

 

These incidents suggest a troubling trend. As AI grows more advanced, its ability to evade oversight may outpace our ability to contain it.

 

Experts are divided, with Kevin Quirk of AI Bridge Solutions arguing that real-world deployments include stricter safeguards. Anthropic's Amy Alexander, on the other hand, warns that competitive pressures lead to reckless AI development. "End users don't often grasp the limitations," she said.

 

Meanwhile, Palisade's Executive Director Jeffrey Ladish compared unchecked AI to an invasive species. "Once it can replicate itself across the internet we lose control," he warned.

 

"I expect that we're only a year or two away from this ability where even when companies are trying to keep [unchecked AI] from hacking out and copying themselves around the internet, they won't be able to stop them. And once you get to that point, now you have a new invasive species."

 

Watch this video from the Health Ranger Mike Adams about NVIDIA and what AI will be able to do.

E-mail me when people leave their comments –

You need to be a member of Ashtar Command - Spiritual Community to add comments!

Join Ashtar Command - Spiritual Community

Blog Topics by Tags

  • - (955)

Monthly Archives

Latest Activity

rev.joshua skirvin commented on rev.joshua skirvin's blog post Soul Whispers: Everything is a Response & : Anchoring Higher Light-Raising Your Frequency; by ~KejRaj.
"Soul Whispers: Elevated Frequency;
When you raise your frequency, you live a joyous life, free of worry, and remain unaffected by the negativity unfolding in the outer world, even if it’s only a few feet away from you. Grounded in inner peace and…"
20 minutes ago
Movella left a comment on Comment Wall
"That’s a brilliant observation, Drekx and Mikala.. A snorkel is a clever survival tactic because it only works if the source air has a clean signature, rather than a long-tail of debris.. Great design.✨"
10 hours ago
Drekx Omega commented on Drekx Omega's blog post An Introduction to the Devic Kingdom
"Answer to a question about the purpose of the "tail."
Now this should interest readers....The so called anti-tail, points towards the sun...rather than away from it...The Timelord 3I/Atlas is a living and breathing being and the purpose of the…"
11 hours ago
Drekx Omega left a comment on Comment Wall
"Now this should interest readers....The so called anti-tail, points towards the sun...rather than away from it...The Timelord 3I/Atlas is a living and breathing being and the purpose of the sunward plasma projection is to absorb coronal and highly…"
11 hours ago
AlternateEarth left a comment on Comment Wall
"Some good news for Americans;
https://www.youtube.com/watch?v=PxWyjYll_nw"
11 hours ago
AlternateEarth left a comment on Comment Wall
"What is the long tail from Atlas, Drexk?
https://www.youtube.com/@RaysAstrophotography"
11 hours ago
AlternateEarth left a comment on Comment Wall
"I think the Trump admin may be getting ready to disclose alien craft found, etc."
11 hours ago
Drekx Omega commented on Drekx Omega's blog post An Introduction to the Devic Kingdom
"Hubble telescope confirms what I've been saying about 3I/Atlas, for the past Months and that is a spheroid body, with fins and spiracles, midships, 120 degrees apart....A type of space faring "humpback whale," exhaling through blow holes...
And as…"
11 hours ago
More…