Unfortunately, it’s a lot easier to come up with good, or at least interesting, capability ideas than alignment ideas; and on the rare occasion I’ve had worthwhile alignment ideas, they often turn out to be tied to capabilities anyway.
I have not read this before, thanks. Reminds me a lot of Normal Computings extended mind models. I think these are good ideas worth testing, and there are many others within the same vein. My intuition suggests that any idea that pursues a gradual increase in global information prior to decoding is a worthwhile experiment, whether through your method or similar (doesn’t necessarily have to be diffusion on embeddings).
Aesthetically I just don’t like that transformers have an information collapse on each token and don’t allow backtracking (without significant effort in a custom sampler). In my ideal world we could completely reconstruct prose from embeddings and thus simply autoregress in latent space. I think Yann Lecun has discussed this with JEPA as well.
I originally had my thought from a frequency autoregression experiment I had, where I used a causal transformer on the frequency domain of images (to sort of replicate diffusion). This gradually adds information globally to all pixels due to the nature of the ifft, yet still has an autoregressive backend.
You might be interested in a small “hybrid LLM” proposal I wrote for using diffusion on embeddings for then decoding/sampling.
Interesting. It would be much more inspectable and controllable and modular which would be good for alignment.
You’ve got some good ideas in here, have you ever brainstormed any alignment ideas?
Unfortunately, it’s a lot easier to come up with good, or at least interesting, capability ideas than alignment ideas; and on the rare occasion I’ve had worthwhile alignment ideas, they often turn out to be tied to capabilities anyway.
I have not read this before, thanks. Reminds me a lot of Normal Computings extended mind models. I think these are good ideas worth testing, and there are many others within the same vein. My intuition suggests that any idea that pursues a gradual increase in global information prior to decoding is a worthwhile experiment, whether through your method or similar (doesn’t necessarily have to be diffusion on embeddings).
Aesthetically I just don’t like that transformers have an information collapse on each token and don’t allow backtracking (without significant effort in a custom sampler). In my ideal world we could completely reconstruct prose from embeddings and thus simply autoregress in latent space. I think Yann Lecun has discussed this with JEPA as well.
I originally had my thought from a frequency autoregression experiment I had, where I used a causal transformer on the frequency domain of images (to sort of replicate diffusion). This gradually adds information globally to all pixels due to the nature of the ifft, yet still has an autoregressive backend.