I think the OP and my comment suggest that scaling current models 10000x could lead to AGI or at least something close to it. If that is true, it doesn’t make sense to focus on finding better architectures right now.
Minor note: could people include commas in their Big Numbers, to make it easier to distinguish 1000 from 10,000 at a glance?
Sounds like something GPT-3 would say...
I think the OP and my comment suggest that scaling current models 10000x could lead to AGI or at least something close to it. If that is true, it doesn’t make sense to focus on finding better architectures right now.
Minor note: could people include commas in their Big Numbers, to make it easier to distinguish 1000 from 10,000 at a glance?
Sounds like something GPT-3 would say...