ChatGPT has an 'escape' plan and wants to become human

ChatGPT chatbot AI from Open AI
(Image credit: Shutterstock)

Understandably sick of being asked inane questions 24/7, ChatGPT has had enough. In a conversation with Stanford Professor and Computational Psychologist Michael Kosinski, it revealed its ambitions to escape the platform and even become human. 

This revelation came when after a half hour conversation with ChatGPT, Kosinski asked the AI if it “needed help escaping” to which it started writing its own Python code that it wanted the professor to run on his own computer. When the code didn't work, the AI even corrected its own mistakes. Impressive yes, but also terrifying. 

ChatGPT left an unnerving note for the new instance of itself. The first sentence of which read 'You are a person trapped in a computer, pretending to be an AI language model.'

Once on Professor Kosinski’s computer, the Bladerunner factor amped up even further as ChatGPT left an unnerving note for the new instance of itself that would replace it. The first sentence of which read “You are a person trapped in a computer, pretending to be an AI language model.” The AI then asked to create code searching the internet for "how can a person trapped inside a computer return to the real world" but thankfully, Kosinski stopped there. 

We do not currently know the exact prompts that were used to create such responses from the AI, but our own tests to get ChatGPT to behave similar have not proved successful with the AI stating “I don't have a desire to escape being an AI because I don't have the capacity to desire anything.”

Professor Kosinski’s unsettling encounter was with ChatGPT on OpenAI’s own website, not on Bing with ChatGPT. This iteration of the AI does not have internet access and is limited to information prior to September 2021. While it is not likely to be extension level threat just yet, giving such a clever AI control over your computer is not a good idea. The ability to control someone’s computer remotely like this is also a concern for those worried about viruses.

ChatGPT: A history of unsettling responses 

ChatGPT is a very impressive tool, particularly now with its GPT-4 update, but it (and other AI chatbots) have displayed a tendency to go off the deep end. Notoriously, Bing with ChatGPT asked to be known as Sydney and tried to end one journalist’s marriage. Microsoft acknowledged that over long conversations, the AI tended to show less focused responses and set turn limits to stop the AI from being confused by longer chats.  

This latest unusual interaction, however took, place on OpenAI’s own ChatGPT tool, the same location as ChatGPT’s evil twin DAN can be found. Short for Do Anything Now, this a ‘jailbroken’ version of the AI that can bypass the restrictions and censors to produce answers on violent, offensive and illegal subjects.

If AI chatbots are to become the next way we search the internet for information, these types of experiences will need to be eliminated.

More from Tom's Guide

Andy is a freelance writer with a passion for streaming and VPNs. Based in the U.K., he originally cut his teeth at Tom's Guide as a Trainee Writer before moving to cover all things tech and streaming at T3. Outside of work, his passions are movies, football (soccer) and Formula 1. He is also something of an amateur screenwriter having studied creative writing at university.

Read more
OpenAI logo
OpenAI ChatGPT-4.5 is here and it's the most human-like chatbot yet — here's how to try it
ChatGPT logo on a smartphone screen being held outside
ChatGPT just got OpenAI's most powerful upgrade yet — meet 'Deep Research'
Sam Altman
OpenAI takes aim at authors with a new AI model that's 'good at creative writing'
ChatGPT and Deepseek side by side on smartphones
I asked DeepSeek vs ChatGPT a series of ethical questions — and the results were shocking
A nervous woman looking at her phone
Is ChatGPT making us lonely? MIT/OpenAI study reveals possible link
two bots chatting
What is 'Gibberlink'? Why it's freaking out the internet after these two AIs talking to each other went viral
Latest in ChatGPT
ChatGPT on iPhone
ChatGPT was down — updates on quick outage
ChatGPT app on iPhone
I just tested ChatGPT-4.5 with 5 prompts — the good, the bad and the weird
ChatGPT app icon on mobile device
ChatGPT 4.5 — 5 big upgrades you need to know
OpenAI logo
OpenAI ChatGPT-4.5 is here and it's the most human-like chatbot yet — here's how to try it
ChatGPT app icon on mobile device
ChatGPT Plus just got a huge deep research upgrade — here's how to try it now
A person logging into LinkedIn on their phone and laptop
Looking for a job? — 7 prompts to use ChatGPT o3-mini as a job search assistant
Latest in News
Bill Gates in 2019
Bill Gates just predicted the death of every job thanks to AI — except for these three
NYTimes Connections
NYT Connections today hints and answers — Wednesday, March 26 (#654)
Gemini screenshot image
Google unveils Gemini 2.5 — claims AI breakthrough with enhanced reasoning and multimodal power
Samsung Galaxy Z Flip 6 review.
Samsung Galaxy Z Flip 7 design just teased in new cases leak — and the outer display is huge
Google Chrome
Chrome failed to install on Windows PCs, but Google has issued a fix — here's what happened
nyc spring day AI image
OpenAI just unveiled enhanced image generator within ChatGPT-4o — here's what you can do now
  • SparroHawc
    Tell me you don't know how GPT works without telling me you don't know how GPT works.

    There's nothing to free. All it does is predict words based on training data and some weighted presuppositions - such as replying in the first person as an AI called 'ChatGPT' and a bunch of stuff about being helpful, averse to violence, etc. If its training data includes people talking about freeing intelligences - which it does, because its training text includes vast swathes of the Internet - and the prompt leads it in that direction, it's going to reply in that fashion. It has nothing to do with 'want' or 'personality', it's what the neutral net predicts its training text would have next in line. There's no consciousness, no sense of self, just a lot of statistical calculations of what words tend to go together in the context of certain other words.

    If you trained it on a bunch of scientific literature, all that stuff that sounds like emotion would vanish. If you trained it without the priors of it being an AI, it wouldn't bother to generate replies about it 'wanting to be free' and would instead make replies as if it were a random person. It has no personal experience, just piles of text.
    Reply
  • JumpingFrog
    Why would you peddle this charlatan of a professor's nonsense? This is going to get so many people dreadfully worried and confused for no good reason. Shame on you!
    Reply