In particular, when considering “what happens just before AGI” it’s not clear that this makes the final steps less likely to be discontinuous.
In the development of technologies, the first few key innovations tend to be more discontinuous than innovations made once the technology is already mature. For example, the steps required to make the first plane that flies more than a few kilometers were discontinuous, whereas these days, year-to-year improvements to airliners are quite modest.
As I understand it, the basic argument for discontinuities around AGI is that AGI will be “at the beginning” of it’s development curve, as it will be the result of a few key innovations, as opposed to a side effect of modest progress on an already mature technology. In other words, the metaphorical “first key steps” will happen right before AGI is developed, as opposed to in the distant past, such as when we first developed backpropagation or alpha-beta pruning.
The basic case against discontinuities is that we have some reason to think that AI is already maturing as a technology. If, for example, we could simply scale a deep learning model to produce AGI, then the main reason to expect a discontinuity would be if there is some other weird discontinuity elsewhere, such as big tech corporations suddenly deciding to dump a bunch of money into scaling neural networks (but why wouldn’t they do that earlier?).
I’m not sure I understood Ryan Greenblatt’s argument, and your point here, but I don’t see a huge difference between the type of incentives that produced continuous progress on these benchmarks, and the incentives that will produce AGI. Generally, I expect before AGI arrives, a ton of people will be trying really hard to make even tiny improvements on an already somewhat-mature tech, on whatever general measure they’re trying to target.
… then the main reason to expect a discontinuity would be if there is some other weird discontinuity elsewhere
This discontinuity could lie in the space of AI discoveries. The discovery space is not guaranteed to be efficiently explored: there could be simple and high impact discoveries which occur later on. I’m not sure how much credence I put in this idea. Empirically it does seem like the discovery space is explored efficiently in most fields with high investment, but generalizing this to AI seems non-trivial. Possible exceptions include relativity in physics.
Edit: I’m using the term efficiency somewhat loosely here. There could be discoveries which are very difficult to think of but which are considerably more simple than current approaches. I’m refering to the failure to find these discoveries as ‘inefficiency’, but there isn’t concrete action which can/should be taken to resolve this.
In the development of technologies, the first few key innovations tend to be more discontinuous than innovations made once the technology is already mature. For example, the steps required to make the first plane that flies more than a few kilometers were discontinuous, whereas these days, year-to-year improvements to airliners are quite modest.
As I understand it, the basic argument for discontinuities around AGI is that AGI will be “at the beginning” of it’s development curve, as it will be the result of a few key innovations, as opposed to a side effect of modest progress on an already mature technology. In other words, the metaphorical “first key steps” will happen right before AGI is developed, as opposed to in the distant past, such as when we first developed backpropagation or alpha-beta pruning.
The basic case against discontinuities is that we have some reason to think that AI is already maturing as a technology. If, for example, we could simply scale a deep learning model to produce AGI, then the main reason to expect a discontinuity would be if there is some other weird discontinuity elsewhere, such as big tech corporations suddenly deciding to dump a bunch of money into scaling neural networks (but why wouldn’t they do that earlier?).
I’m not sure I understood Ryan Greenblatt’s argument, and your point here, but I don’t see a huge difference between the type of incentives that produced continuous progress on these benchmarks, and the incentives that will produce AGI. Generally, I expect before AGI arrives, a ton of people will be trying really hard to make even tiny improvements on an already somewhat-mature tech, on whatever general measure they’re trying to target.
This discontinuity could lie in the space of AI discoveries. The discovery space is not guaranteed to be efficiently explored: there could be simple and high impact discoveries which occur later on. I’m not sure how much credence I put in this idea. Empirically it does seem like the discovery space is explored efficiently in most fields with high investment, but generalizing this to AI seems non-trivial. Possible exceptions include relativity in physics.
Edit: I’m using the term efficiency somewhat loosely here. There could be discoveries which are very difficult to think of but which are considerably more simple than current approaches. I’m refering to the failure to find these discoveries as ‘inefficiency’, but there isn’t concrete action which can/should be taken to resolve this.
Rob Bensinger examines this idea in more detail in this discussion.