Almost as soon as humans dreamed up the idea of robots, futurists were imagining worst-case scenarios where they would run amok and ruin our lives.
Way back in 1941, for example, Superman was valiantly fighting pitched battles with giant robots in the pages of comic books. More recently, we’ve had movies like ‘Blade Runner’ or ‘Terminator’ paint a bleak future where sentient robots go rogue and rage against the humans. Those stories weren’t just an artifact of the 80s either, as evidenced by the success of ‘I, Robot’.
These worries fed into already-existing fears of workplace automation putting workers out of jobs, and the more modern version of the same concern — where AI is doing jobs we used to pay people for.
But surely getting worked up about sentient robots going berserk is as ridiculous as being worried about waking up in the world pictured in classic monster films or, say, ‘The Handmaid’s Tale’. Right?
Maybe. But then again… maybe not.
There are two recent news stories that are just kind interesting on their own. But taken together — especially when our top inventors are proposing a world with driverless cars and android workers will be everywhere, doing the jobs that are difficult, dangerous, or just plain menial, they present a disturbing vision of what could go wrong, and how difficult it would be to fix it.
The Funny Part
Since the other part is a little more serious and technical, let’s begin with this clip of a catastrophic failure of a Chinese-manufactured humanoid robot.
This week’s episode is out! Surveillance footage showing a Unitree H1 robot, secured by a crane, suddenly flailing its arms out of control has been spreading online. In the video, a frightened worker can be seen trying to turn off the power. Topics include ads to children, cost… pic.twitter.com/xZYBkdcRgN
— バイリンガルニュース (@Bilingual_News) May 15, 2025
Fortunately, it was secured to a chain and staff there were able to hit the ‘kill switch’ before anyone got hurt.
We are told that there was some faulty code in its programming, and that led to the dangerously erratic behavior.
Keep that incident in the back of your mind as we move to the more serious part… a study in which we learn that AI is behaving less like an obedient machine, and more like … well, judge for yourself.
It’s writing code in its own ‘brain’ to override the shutdown mechanism?
Even before we look further, it doesn’t take a lot of creativity to imagine what might have gone wrong if the robot on the crane had no ‘kill switch’. Especially in a real-world situation where it wasn’t tethered to a crane.
Claude Opus 4 is part of a Bezos-backed AI enterprise, and recent revelations have shown just how great the potential for hazard can be.
Axios summarized some of the more disturbing parts of what the study revealed…
Driving the news: Anthropic on Thursday announced two versions of its Claude 4 family of models, including Claude 4 Opus, which the company says is capable of working for hours on end autonomously on a task without losing focus.
Anthropic considers the new Opus model to be so powerful that, for the first time, it’s classifying it as a Level 3 on the company’s four-point scale, meaning it poses “significantly higher risk.”
[…]
In one scenario highlighted in Opus 4’s 120-page “system card,” the model was given access to fictional emails about its creators and told that the system was going to be replaced.
On multiple occasions it attempted to blackmail the engineer about an affair mentioned in the emails in order to avoid being replaced, although it did start with less drastic efforts.
Meanwhile, an outside group found that an early version of Opus 4 schemed and deceived more than any frontier model it had encountered and recommended against releasing that version internally or externally.
“We found instances of the model attempting to write self-propagating worms, fabricating legal documentation, and leaving hidden notes to future instances of itself all in an effort to undermine its developers’ intentions,” Apollo Research said in notes included as part of Anthropic’s safety report for Opus 4.
Of course, there was more revealed than a little news blurb like that could cover.
Tech Crunch, for example, went into more detail about the forms of deception:
Per Anthropic’s report, Apollo observed examples of the early Opus 4 attempting to write self-propagating viruses, fabricating legal documentation, and leaving hidden notes to future instances of itself — all in an effort to undermine its developers’ intentions.
Meanwhile, we’ve got a guy on YouTube running a head-to-head battle between two AI platforms having them solve a complex logic problem, where he discovered that Opus ‘social pressure override’ was responsible for Opus asserting as true something that it had only a human’s untested say-so for accepting (namely that there were multiple solutions to a logic problem that has only a single resolution). It went on to claim there were 3 solutions, and then failed to demonstrate what they were. The fun part begins around the 10 minute mark.
These two stories, taken together seemed to set up a potentially volatile scenario. So went a step further, and asked Grok to consider these two stories — including the deception — and imagine a worst-case scenario.
I allowed for a world in which some of the technologies our thought leaders are already developing and dreaming of mass-producing become commonplace, in a world that bears some resemblance to the one imagined in I, Robot.
Here’s what Grok came back with.
Obviously, a worst-case scenario is exactly that.
And there are layers of layers of safety protocols that would make such a scenario more science fiction than science fact. The beserk robot was in a testing facility, secured to a crane when it went haywire. The AI issues were reported by an organization whose job is to identify them.
That being said, the whole point of troubleshooting is to imagine what can go wrong and look for ways to mitigate the known or imagined problems. We took known issues, and turned them up to 11, and placed them in something like the world that tech visionaries already want our children (or possibly even us) to live in.
AI has been trained to do tasks that would have beggared the imagination just a few years ago. Its growth shows no signs of slowing. Solving the problem of founding ‘morally flexible’ in a framework that prioritizes integrity AI is something we will want to get right sooner, rather than later.
When even AI is capable of imagining (an admittedly) worst-case scenario of Skynet proportions where it postulates a potential war of machines against men with malicious code and mass fatalities?
Is such a scenario melodramatic? Of course. But like any other use of imagery, including the Cold War what-if film that made President Reagan so cautious about Mutually Assured Destruction, or Al Gore’s environmental equivalent …jarring imagery can serve a purpose, and get the wheels turning on a subject that didn’t gain much interest.
Maybe now is the time to start taking more interest in what questions we are really asking about our safety protocols… especially since the nature of AI is for it to grow and advance in ways that even its creators don’t truly understand.
The post SKYNET: Two News Stories That Show Worst Case Scenario Isn’t (Quite) As Far-Fetched As We Think appeared first on Clash Daily.