Thanks for the catch, I deleted “Note that the hidden dimen”. Transformers do blow up the hidden dimension, but that’s not very relevant here—they have many more neurons than residual stream dimensions, and they have many more features than neurons (as shown in the recent Anthropic paper)
Is the point that transformer MLPs blow up the hidden dimension in the middle?
Thanks for the catch, I deleted “Note that the hidden dimen”. Transformers do blow up the hidden dimension, but that’s not very relevant here—they have many more neurons than residual stream dimensions, and they have many more features than neurons (as shown in the recent Anthropic paper)
To clarify, I thought it was about superposition happening inside the projection afterwards.