Yeah, I think that reasoning, along with various other AGI prerequisites, requires an algorithm that does probabilistic programming / analysis-by-synthesis during deployment. And I think that trained Transformer models don’t do that, no matter what their size and parameters are. I guess I should write a post about why I think that—it’s a bit of a hazy tangle of ideas in my mind right now. :-)
(I’m more-or-less saying the interpretation you disagree with in your second-to-last paragraph.)
Got it!
Yeah, I think that reasoning, along with various other AGI prerequisites, requires an algorithm that does probabilistic programming / analysis-by-synthesis during deployment. And I think that trained Transformer models don’t do that, no matter what their size and parameters are. I guess I should write a post about why I think that—it’s a bit of a hazy tangle of ideas in my mind right now. :-)
(I’m more-or-less saying the interpretation you disagree with in your second-to-last paragraph.)
Thanks again for explaining!