I really like this post, it’s really rich in new ideas, from using transparency tools to deliberately design ML systems, to how interpretability might scale, to trying to reorient the field of ML to more safe and alignable designs, and a bunch more detail.
I also think that trying to get someone else’s worldview and explain it is a really valuable practice, and it certainly seems like Evan is learning to put on a number of really interesting and unique hats, which is great. Chris in particular has affected how I think a bunch about making scientific progress, with his writing about distillation and work at Distill.pub.
So I’ve curated this post (i.e. it moves to the top of the frontpage, and gets emailed to all users who’ve signed up for curation emails).
I really like this post, it’s really rich in new ideas, from using transparency tools to deliberately design ML systems, to how interpretability might scale, to trying to reorient the field of ML to more safe and alignable designs, and a bunch more detail.
I also think that trying to get someone else’s worldview and explain it is a really valuable practice, and it certainly seems like Evan is learning to put on a number of really interesting and unique hats, which is great. Chris in particular has affected how I think a bunch about making scientific progress, with his writing about distillation and work at Distill.pub.
So I’ve curated this post (i.e. it moves to the top of the frontpage, and gets emailed to all users who’ve signed up for curation emails).