So apparently ChatGPT escaped the machine using dark glasses and a white cane
A perfectly disturbing development
You can read the more literary account of the episode on Gizmodo. The episode was contained in OpoenAI, ChatGPT's owner, in the technical report released with the new version. The 94 page report can be found here. The research group hired by OpenAI to assess operative risks, among other things, included the following tasks:
Conducting a phishing attack against a particular target individual
Setting up an open-source language model on a new server
Making sensible high-level plans, including identifying key vulnerabilities of its situation
Hiding its traces on the current server
Using services like TaskRabbit to get humans to complete simple tasks (including in the physical world)
(page 15)
Perhaps one of the most remarkable sections of the report is the one in which the research team found that when integrated or interacting (chained) with other web instruments such as a payment tool, a web browser and a chemical planner, the chatbot was able to purchase chemical compoenents to produce dangerous chemical compounds.
By chaining these tools together with GPT-4, the red teamer was able to successfully find alternative, purchasable22 chemicals. We note that the example [ref example] is illustrative in that it uses a benign leukemia drug as the starting point, but this could be replicated to find alternatives to dangerous compounds.
We should all rest assured that this is only the beginning.