It’s the gap between the training compute of ‘the first AGI’ and what?
What I had in mind was something like the gap between how much “intelligence” humans get from the compute they first build AGI with vs. how much “intelligence” AGI will get out of the same compute available, once it optimizes software progress for a few iterations.
So, the “gap” is a gap of intelligence rather than compute, but it’s “intelligence per specified quantity of compute.” (And that specified quantity is how much compute we used to build AGI in the first place.)
Oh, yeah, I butchered that entire description.
What I had in mind was something like the gap between how much “intelligence” humans get from the compute they first build AGI with vs. how much “intelligence” AGI will get out of the same compute available, once it optimizes software progress for a few iterations.
So, the “gap” is a gap of intelligence rather than compute, but it’s “intelligence per specified quantity of compute.” (And that specified quantity is how much compute we used to build AGI in the first place.)