The case in my mind for preferring to elicit and solve problems at scale rather than in toy demos (when that’s possible) is pretty broad and outside-view, but I’d nonetheless bet on it: I think a general bias toward wanting to “practice something as close to the real thing as possible” is likely to be productive. In terms of the more specific benefits I laid out in this section, I think that toy demos are less likely to have the first and second benefits (“Practical know-how and infrastructure” and “Better AI situation in the run-up to superintelligence”), and I think they may miss some ways to get the third benefit (“Discovering or verifying a long-term solution”) because some viable long-term solutions may depend on some details about how large models tend to behave.
I do agree that working with larger models is more expensive and time-consuming, and sometimes it makes sense to work in a toy environment instead, but other things being equal I think it’s more likely that demos done at scale will continue to work for superintelligent systems, so it’s exciting that this is starting to become practical.
Thanks for the response! I see the approaches as more complimentary. Again, I think this is in keeping with standard/good ML practice.
A prototypical ML paper might first describe a motivating intuition, then formalize it via a formal model and demonstrate the intuition in that model (empirically or theoretically), then finally show the effect on real data.
The problem with only doing the real data (i.e. at scale) experiments is that it can be hard to isolate the phenomena you wish to study. And so a positive result does less to confirm the motivating intuition, as there are many other factors as play that might be responsible. We’ve seen this happen rather a lot in Deep Learning and Deep RL, in part because of the focus on empirical performance over a more scientific approach.
The case in my mind for preferring to elicit and solve problems at scale rather than in toy demos (when that’s possible) is pretty broad and outside-view, but I’d nonetheless bet on it: I think a general bias toward wanting to “practice something as close to the real thing as possible” is likely to be productive. In terms of the more specific benefits I laid out in this section, I think that toy demos are less likely to have the first and second benefits (“Practical know-how and infrastructure” and “Better AI situation in the run-up to superintelligence”), and I think they may miss some ways to get the third benefit (“Discovering or verifying a long-term solution”) because some viable long-term solutions may depend on some details about how large models tend to behave.
I do agree that working with larger models is more expensive and time-consuming, and sometimes it makes sense to work in a toy environment instead, but other things being equal I think it’s more likely that demos done at scale will continue to work for superintelligent systems, so it’s exciting that this is starting to become practical.
Thanks for the response!
I see the approaches as more complimentary.
Again, I think this is in keeping with standard/good ML practice.
A prototypical ML paper might first describe a motivating intuition, then formalize it via a formal model and demonstrate the intuition in that model (empirically or theoretically), then finally show the effect on real data.
The problem with only doing the real data (i.e. at scale) experiments is that it can be hard to isolate the phenomena you wish to study. And so a positive result does less to confirm the motivating intuition, as there are many other factors as play that might be responsible. We’ve seen this happen rather a lot in Deep Learning and Deep RL, in part because of the focus on empirical performance over a more scientific approach.