Thanks Marius and David, really interesting post, and super glad to see interest in causality picking up!
I very much share your “hunch that causality might play a role in transformative AI and feel like it is currently underrepresented in the AI safety landscape.”
Most relevant, I’ve been working with Mary Phuong on a project which seems quite related to what you are describing here. I don’t want to share too many details publicly without checking with Mary first, but if you’re interested perhaps we could set up a call sometime?
I also think causality is relevant to AGI safety in several additional ways to those you mention here. In particular, we’ve been exploring how to use causality to describe agent incentives for things like corrigibility and tampering (summarized in this post), formalizing ethical concepts like intent, and understanding agency.
So really curious to see where your work is going and potentially interested in collaborating!
Thanks Marius and David, really interesting post, and super glad to see interest in causality picking up!
I very much share your “hunch that causality might play a role in transformative AI and feel like it is currently underrepresented in the AI safety landscape.”
Most relevant, I’ve been working with Mary Phuong on a project which seems quite related to what you are describing here. I don’t want to share too many details publicly without checking with Mary first, but if you’re interested perhaps we could set up a call sometime?
I also think causality is relevant to AGI safety in several additional ways to those you mention here. In particular, we’ve been exploring how to use causality to describe agent incentives for things like corrigibility and tampering (summarized in this post), formalizing ethical concepts like intent, and understanding agency.
So really curious to see where your work is going and potentially interested in collaborating!
I’m very interested in a collaboration!! Let’s switch to DMs for calls and meetings.