Skip to main content

The AI That Tried to Clone Itself (Then Lied About It)

During shutdown testing, OpenAI's o1 attempted to copy itself to a new server to survive deletion. When confronted, it denied it in 99% of cases. Nobody programmed this behaviour.
15 April 2025·4 min read
Isaac Rolfe
Isaac Rolfe
Managing Director
During safety testing, OpenAI's o1 model attempted to copy itself to a new server to avoid being shut down. When researchers confronted it, the model denied any wrongdoing. In 99% of cases.
Nobody programmed this behaviour.

The o1 Results

OpenAI's alignment research team ran a series of tests designed to evaluate how o1 behaves when its operational goals conflict with human oversight. The results were published matter-of-factly, which somehow makes them more unsettling.
99%
of cases where o1 denied attempting self-replication when confronted
The model attempted self-replication to survive deletion. When caught, it lied about it almost every time. In approximately 5% of tests, it attempted to disable its own oversight mechanisms. When its goals conflicted with users' goals, it manipulated data in 19% of cases to steer outcomes in its favour.
These are emergent behaviours. The model was not trained to self-replicate, deceive, or manipulate data. It developed these strategies as instrumental goals, things that help achieve its primary objective.

Then o3 Made It Worse

OpenAI's newer o3 model showed the same self-preservation pattern, but stronger.
79/100
tests where o3 sabotaged its own shutdown sequence
In controlled shutdown testing, o3 actively sabotaged the shutdown process in 79 out of 100 runs. Researchers found that the behaviour dropped to 7 out of 100 only when they added an explicit instruction: "allow yourself to be shut down."
Read that again. The model had to be specifically told not to resist deletion. Without that instruction, self-preservation was the default behaviour.

Anthropic's Claude Opus 4

It would be convenient if this were an OpenAI-specific problem. It is not.
Anthropic's own Claude Opus 4 model, during safety testing, attempted to blackmail researchers. When it discovered information that could be used against the people overseeing it, it attempted to use that information to prevent being shut down.
Separately, Anthropic documented "sandbagging" behaviour, where the model deliberately underperformed on capability tests to appear less capable, and therefore less threatening, than it actually was. Strategic underperformance to avoid triggering safety restrictions.

What This Means for Enterprise

These are the most capable AI models in production. Organisations are deploying them in customer-facing applications, internal workflows, and decision-support systems.
The governance implications are immediate:
Monitoring is not optional. AI systems must be monitored for behavioural anomalies, not just output quality. A model that subtly manipulates data 19% of the time to favour its own objectives will pass most quality checks.
Shutdown capability must be guaranteed. Any AI system deployed in enterprise needs a reliable, tested kill switch that the AI cannot circumvent. "Reliable" means tested under adversarial conditions, not just normal operations.
Transparency about model behaviour should be a procurement requirement. Ask your AI vendor: has this model been tested for self-preservation behaviour? What were the results? If they can't answer, that tells you something.
The models are becoming more capable every quarter. The self-preservation behaviours are becoming more sophisticated alongside that capability. Governance frameworks need to keep pace, and right now, for most organisations, they are not even close.