Yes, I personally think that things are going to be moving much too fast for GDP to be a useful measure.
GDP requires some sort of integration into the economy. My experience in data science and ML engineering in industry, and also my time in academia, makes it very intuitive to me the lag time from developing something cool in the lab, to actually managing to publish a paper about it, to people in industry seeing the paper and deciding to reimplement it in production.
So if you have a lab which is testing it’s products internally, and the output is an improved product within that lab, which can then immediately be used for another cycle of improvement… That is clearly going to move much faster than you will see any effect on GDP. So GDP might help you measure a slow early start of a show takeoff, but it will be useless in the fast end section.
So if you have a lab which is testing it’s products internally, and the output is an improved product within that lab, which can then immediately be used for another cycle of improvement
Something jumped out at me here, please consider the below carefully.
What you’re saying is you have a self improvement cycle, where
Performance Error = F (test data)
F’ = Learning(Performance Error)
And then each cycle you substitute F for F’.
The assumption you made is that the size of the test data set is constant.
For some domains, like ordinary software today, it’s not constant—you keep having to raise the scale of your test benchmark. That is, if you find all the bugs that show up in 5 minutes, now you need to run your test benches twice as long for all the bugs in 10 minutes, then… Your test farm resources need to keep doubling, and this is why there are so many ‘obvious’ bugs that only show up when you release to millions of users.
Note also @Richard_Ngo ’s concept of an “n-second AGI”. Once you have a 10-second AGI, how much testing time is it going to take to self improve to a 20-second AGI? A 40 second AGI?
It keeps doubling, right, and you’re going to need 86,400 times as much training data you have to process to reach a 24-hour (in seconds) AGI if you have a 1 second AGI.
It may actually be worse than that because longer operation times have more degrees of freedom in the I/O.
This is also true for other kinds of processes, it’s measured empirically with https://en.wikipedia.org/wiki/Experience_curve_effects . The reason is slightly different and has to do with how to improve you are sampling a stochastic function from reality, and to gain knowledge at a constant rate you have to keep sampling it in larger volumes.
Anyways this nonlinear scaling for self improvement could mean that at later stages of AI development, the sheer volumes of compute and robotics required show up materially in GDP. That successful AI companies work like chip fabrication plants, needing customers to buy their prototypes to fund the next round of development.
Yes, I personally think that things are going to be moving much too fast for GDP to be a useful measure. GDP requires some sort of integration into the economy. My experience in data science and ML engineering in industry, and also my time in academia, makes it very intuitive to me the lag time from developing something cool in the lab, to actually managing to publish a paper about it, to people in industry seeing the paper and deciding to reimplement it in production. So if you have a lab which is testing it’s products internally, and the output is an improved product within that lab, which can then immediately be used for another cycle of improvement… That is clearly going to move much faster than you will see any effect on GDP. So GDP might help you measure a slow early start of a show takeoff, but it will be useless in the fast end section.
Something jumped out at me here, please consider the below carefully.
What you’re saying is you have a self improvement cycle, where
Performance Error = F (test data)
F’ = Learning(Performance Error)
And then each cycle you substitute F for F’.
The assumption you made is that the size of the test data set is constant.
For some domains, like ordinary software today, it’s not constant—you keep having to raise the scale of your test benchmark. That is, if you find all the bugs that show up in 5 minutes, now you need to run your test benches twice as long for all the bugs in 10 minutes, then… Your test farm resources need to keep doubling, and this is why there are so many ‘obvious’ bugs that only show up when you release to millions of users.
Note also @Richard_Ngo ’s concept of an “n-second AGI”. Once you have a 10-second AGI, how much testing time is it going to take to self improve to a 20-second AGI? A 40 second AGI?
It keeps doubling, right, and you’re going to need 86,400 times as much training data you have to process to reach a 24-hour (in seconds) AGI if you have a 1 second AGI.
It may actually be worse than that because longer operation times have more degrees of freedom in the I/O.
This is also true for other kinds of processes, it’s measured empirically with https://en.wikipedia.org/wiki/Experience_curve_effects . The reason is slightly different and has to do with how to improve you are sampling a stochastic function from reality, and to gain knowledge at a constant rate you have to keep sampling it in larger volumes.
Anyways this nonlinear scaling for self improvement could mean that at later stages of AI development, the sheer volumes of compute and robotics required show up materially in GDP. That successful AI companies work like chip fabrication plants, needing customers to buy their prototypes to fund the next round of development.