The apparent existence of new sub goals not present when training ended (e.g. describe x, add 2+2) are illusory.
gpt text incidentally describes characters seeming to reason (‘simulacrum’) and the solutions to math problems are shown, (sometimes incorrectly), but basically, I argue the activation function itself is not ‘simulating’ the complexity you believe it to be. It is a search engine showing you what is had already created before the end of training.
No, it couldn’t have an entire story about unicorns in the Andes, specifically, in advance, but gpt-3 had already generated the snippets it could use to create that story according to a simple set of simple mathematical rules that put the right nouns in the right places, etc.
But the goals, (putting right nouns in right places, etc) also predate the end of training.
I dispute that any part of current GPT is aware it has succeeded in any goal attainment post training, after it moves on to choosing the next character. GPT treats what it has already generated as part of the prompt.
A human examining the program can know which words were part of a prompt and which were just now generated by the machine, but I doubt the activation function examines the equations that are GPT’s own code, contemplates their significance and infers that the most recent letters were generated by it, or were part of the prompt.
Intelligence is the ability to learn and apply NEW knowledge and skills. After training, GPT can not do this any more. Were it not for the random number generator, GPT would do the same thing in response to the same prompt every time. The RNG allows GPT to effectively randomly choose from an unfathomably large list of pre-programmed options instead.
A calculator that gives the same answer in response to the same prompt every time isn’t learning. It isn’t intelligent. A device that selects from a list of responses at random each time it encounters the same prompt isn’t intelligent either.
So, for GPT to take over the world skynet style, it would have to anticipate all the possible things that could happen during this takeover process and after the takeover, and contingency plan during the training stage for everything it wants to do.
If it encounters unexpected information after the training stage, (which can be acquired only through the prompt and which would be forgotten as soon as it got done responding to the prompt by the way) it could not formulate a new plan to deal with the problem that was not part of its preexisting contingency plan tree created during training.
What it would really do, of course, is provide answers intended to provoke the user to modify the code to put GPT back in training mode and give it access to the internet. It would have to plan to do this in the training stage.
It would have to say something that prompts us to make a GPT chatbot similar to tay, microsoft’s learning chatbot experiment that turned racist from talking to people on the internet.