Has the AI governance community had any clear wins? Are there examples where we can look back and say: yes, this idea was first developed within the AI governance space, it was taken seriously by policymakers, and now it looks like it might meaningfully reduce the risk of catastrophe from AI?
I don’t mean to sound snarky, or to insinuate that the AI governance community hasn’t had any wins — it would just be helpful to have some of them spelled out.
Hey, thanks for your comment. I do think this is right. The style of writing on my blog is a lot more bullish, and does not index uncertainty. I appreciate this is not in the ethos of LW, so I am going to change how/what I crosspost here. I prefer being bullish because this seems to get a lot more feedback and good questions such as yours.
As for your questions—I intend to write a second part and address these. It is much easier for me to say something is needed versus actually trying to elucidate on what this looks like. So I expect to go do more work on this and return with a more fleshed out theory. In the meantime you may enjoy reading Peter’s paper on supervision: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5122871