This post surprised me a lot. It still surprises me a lot, actually. I’ve also linked it a lot of times in the past year.
The concrete context where this post has come up is in things like ML transparency research, as well as lots of theories about what promising approaches to AGI capabilities research are. In particular, there is a frequently recurring question of the type “to what degree do optimization processes like evolution and stochastic gradient descent give rise to understandable modular algorithms?”.
This post surprised me a lot. It still surprises me a lot, actually. I’ve also linked it a lot of times in the past year.
The concrete context where this post has come up is in things like ML transparency research, as well as lots of theories about what promising approaches to AGI capabilities research are. In particular, there is a frequently recurring question of the type “to what degree do optimization processes like evolution and stochastic gradient descent give rise to understandable modular algorithms?”.