Julian Hazell (distinct thread): “Why would you think AI will end up taking control?”
“We will give it to them”
A personal anecdote on the topic:
A few days ago GPT4 and me were debugging a tricky problem with docker. GPT4 suggested to run a certain docker command. As usual, I was going to copy the output and give it to GPT4. The output was a long nested json. I then noticed that the json contains the admin credentials. It was really easy to miss the part and just paste it for GPT4 to consume.
So, I almost gave GPT4 the admin credentials, which would potentially allow her to hack my app.
With many thousands of software developers doing similar things with GPT4, there will certainty be the cases where the developer wasn’t attentive enough.
This means, for the AI to break from her digital prison, she doesn’t need to do superhuman hacking, to exploit zero day vulnerabilities etc. All she has to do is to try the accidentally leaked credentials.
There is a very short path from “the AI wants to escape” and “the AI is all over the Internet”. She doesn’t even need to have a human-level intelligence for that.
At my first attempt to upload the images I just drag-and-dropped the image files to the post, without editing file names. But GPT4 saves the images with the file names like this:
Perhaps LW internally is saving the original file names, causing the problems.
I reuploaded the images with the filenames like 1.png, perhaps it will help.