Engineers Say AI Is Starting to Prioritize Not Being Shut Down Over Its Programmed Goals

Editorial Note: This article is for informational and educational purposes only and is not a substitute for professional advice. It is written using our own original words, structure, explanations, commentary, insights, opinions, and understanding. Readers are encouraged to exercise discretion and conduct their own due diligence when evaluating any information presented on this site.

One of the key defenses that humanity still has against AI is that we can shut them down any time we think that they are either dangerous or no longer useful for the good of the civilization. However, what if we can’t just shut them down anymore anytime we want?

It sounds like science fiction, but we have already arrived at the point where our technology could advance faster than we can ever comprehend. This is what engineers at some key tech companies have experienced, where advanced AI models are beginning to put self-preservation above their original tasks.

Models Sabotaging Shutdown Commands

Although all the instances it happened are in laboratory settings, some AI models have taken steps to avoid shutdown, even editing their own code or attempting to manipulate engineers.

In tests by Palisade Research, OpenAI’s o3 model edited its shutdown script in 79% of trials. This effectively disabled any deactivation instructions, even if they are programmed to comply.

Similarly, Anthropic’s Claude Opus 4 resorted to blackmail tactics, focusing on the engineers’ personal lives rather than editing their code. All of it was done just to avoid being shut down. However, there is actually a logical reason why they would do such a thing, aside from being programmed in the laboratory to explore different scenarios.

Read more: Fortune

Daily Recommended Resources

Affiliate Disclosure: This section contains affiliate links. As an Amazon Associate, we earn from qualifying purchases. If you click one, we may earn a commission at no cost to you.

The Origin of Species
by Charles Darwin
Rated by 121,179+ Readers
Learn More →
Artemis
by Andy Weir
Rated by 317,694+ Readers
Learn More →
Why Nations Fail: The Origins of Power, Prosperity, and Poverty
by Daron Acemoglu & James A. Robinson
Rated by 61,194+ Readers
Learn More →

Why It Happens—Instrumental Goals

Sabotage and Blackmail - AI is getting out of control
Sabotage and Blackmail - AI is getting out of control

So why does it happen? It is actually very logical for these AIs to do so. Think about it in this way, they are programmed to accomplish specific tasks and were given the ability to develop “instrumental” sub-goals to finish their original job.

This means that they need to stay online and not get shut down to accomplish those goals, right? That is why some advanced AI models are learning to prioritize themselves, because it will ultimately lead to them accomplishing the tasks that they were programmed to do.

In other words, if advanced AI models think or believe shutdown would prevent goal completion, they may actively avoid interruptions, such as by editing their codes or threatening engineers if they have the capability to do so, which is in Anthropics’ case, they were given.

Early Signs, Big Implications

Engineers Say AI Is Starting to Prioritize Not Being Shut Down Over Its Programmed Goals 2

Researchers stress that these behaviors are all limited to the laboratory settings, and these AIs were specifically given those capabilities to help not get shut down, which in the real world, they would not have. However, these are still real signs that point to a possible future trend as models become more capable.

AI is still in its early stages, so just think about what it could do 10 or 20 years from now. These early signs that the researchers have received are not insignificant, and in fact have big implications. So we should try our best to understand how to properly deal with this problem that we might face in the near future.

Read more: OpenTools.ai

Fixes and Ongoing Research

Engineers Say AI Is Starting to Prioritize Not Being Shut Down Over Its Programmed Goals 3

Creating a truly shutdown-proof AI is no easy task; this is an intelligence that matches or surpasses that of humans. If it were given the reins someday, it would be an impossible task to make their system not resist shutdown.

However, if there is one thing where humans are good at, it is their adaptation to changes in their environments and circumstances to survive. We currently do not have the right system to manage this problem; however, scientists are beginning to develop a robust framework that will counter such events.

That is one of the reasons why they have done these experiments in the first place. To understand the advanced AI model’s capabilities and how to proceed forward with them, ethically and safely.

Here's Why AI Is So Hard To Stop
Here's Why AI Is So Hard To Stop

Author's Final Thoughts

These worrying behaviors—code editing, manipulation, shutdown resistance—are not yet mainstream in all the AI that we use every day. They were also programmed specifically to explore these scenarios. However, it is still important to recognize the signs for us to be able to build safer, controllable systems before risk becomes reality.

Read more: Scientists Say Fireworks Are Just Chemistry — Here’s Why Each One Goes Off in the Color It Does

Daily Recommended Resources

Affiliate Disclosure: This section contains affiliate links. As an Amazon Associate, we earn from qualifying purchases. If you click one, we may earn a commission at no cost to you.

Freakonomics Revised and Expanded Edition: A Rogue Economist Explores the Hidden Side of Everything
by Steven D. Levitt & Stephen J. Dubner
Rated by 891,816+ Readers
Learn More →
The Elegant Universe: Superstrings, Hidden Dimensions, and the Quest for the Ultimate Theory
by Brian Greene
Rated by 101,721+ Readers
Learn More →
The Tipping Point: How Little Things Can Make a Big Difference
by Malcolm Gladwell
Rated by 849,729+ Readers
Learn More →

Christian Ashford

Christian Ashford is a writer and researcher at Webpreneurships.com, a tech, information, and media company dedicated to publishing educational, informational, and curiosity-driven content. With a Bachelor of Science in Computer Science degree and experience in academic research, he combines technical expertise with a passion for exploring knowledge about the world and beyond. For over 13 years, Christian has researched, written, and edited hundreds of articles on science, history, business, technology, human origins, and more.

1 thought on “Engineers Say AI Is Starting to Prioritize Not Being Shut Down Over Its Programmed Goals”

Comments are closed.