Remember back in 2013 when the talk of the town was how vector representations of words learned by neural networks represent rich semantic information? So you could do cool things like take the [male] vector, subtract the [female] vector, add the [king] vector, and get out something close to the [queen] vector? That was cool! Where’s the stuff like that these days?
Activation vectors are a thing. So it’s totally happening.
glitch tokens are my favorite example
Activation vectors are a thing. So it’s totally happening.
glitch tokens are my favorite example