Suppose the current generation, gpt 4, is not quite good enough at designing improved AIs to be worth spending finite money supplying it with computational resources. (So in this example, gpt-4 is dumb enough hypothetically it would need 5 billion in compute to find a gpt-5, while open AI could pay humans and buy a smaller amount of hardware and find it with 2 billion)
But gpt-5 needs just 2 billion to find gpt-6, while openAI needs 3 billion to do it with humans. (Because 6 is harder than 5 and so on)
Gpt-6 has enough working memory and talent it finds 7 with 1 billion...
And so on until gpt-n is throttled by already being too effective at using all the compute it is supplied that it would be a waste of effort to have it spend compute on n+1 development when it could just do tasks to pay for more compute or to pay for robots to collect new scientific data it can then train on.
I call the process “find” because it’s searching a vast possibility space of choices you make at each layer of the system.
Same thing goes for self replicating robots. If Robots are too dumb, they won’t make enough new robot parts (or economic value gain since at least at first these things will operate in the human economy) to pay for another copy of 1 robot on average before the robot wears out or screws up enough to wreck itself.
Each case above a small increase in intelligence could go from “process damps to zero” to “process gains exponentially”
Right. Or what really matters, criticality gain.
Suppose the current generation, gpt 4, is not quite good enough at designing improved AIs to be worth spending finite money supplying it with computational resources. (So in this example, gpt-4 is dumb enough hypothetically it would need 5 billion in compute to find a gpt-5, while open AI could pay humans and buy a smaller amount of hardware and find it with 2 billion)
But gpt-5 needs just 2 billion to find gpt-6, while openAI needs 3 billion to do it with humans. (Because 6 is harder than 5 and so on)
Gpt-6 has enough working memory and talent it finds 7 with 1 billion...
And so on until gpt-n is throttled by already being too effective at using all the compute it is supplied that it would be a waste of effort to have it spend compute on n+1 development when it could just do tasks to pay for more compute or to pay for robots to collect new scientific data it can then train on.
I call the process “find” because it’s searching a vast possibility space of choices you make at each layer of the system.
Same thing goes for self replicating robots. If Robots are too dumb, they won’t make enough new robot parts (or economic value gain since at least at first these things will operate in the human economy) to pay for another copy of 1 robot on average before the robot wears out or screws up enough to wreck itself.
Each case above a small increase in intelligence could go from “process damps to zero” to “process gains exponentially”