MIRI recently announced a new research agenda focused on “agent foundations”. Yet even the Open Philanthropy Project, made up of people who at least share MIRI’s broad worldview, can’t decide whether that research direction is promising or useless. The Berkeley Center for Human-Compatible AI doesn’t seem to have a specific research agenda beyond Stuart Russell. The AI100 Center at Stanford is just kicking off. That’s it.
There’s also:
MIRI’s Alignment for Advanced Machine Learning Systems agenda
The Concrete Problems agenda by Amodei, Olah and others
Russell’s Research Priorities doc written with Dewey and Tegmark, covers probably more than his CHCAI Centre
Owain Evans, Stuart Armstrong and Eric Drexler at FHI
Paul Christiano’s thinking on AI Control
OpenAI’s safety team in formation
DeepMind’s safety team
(?) Wei Dai’s thinking on metaphilosophy and AI… He occasionally comments e.g. on AgentFoundations
Other Machine Learning researchers, e.g. safe exploration in Deep RL, transparency.
Nitpick:
There’s also:
MIRI’s Alignment for Advanced Machine Learning Systems agenda
The Concrete Problems agenda by Amodei, Olah and others
Russell’s Research Priorities doc written with Dewey and Tegmark, covers probably more than his CHCAI Centre
Owain Evans, Stuart Armstrong and Eric Drexler at FHI
Paul Christiano’s thinking on AI Control
OpenAI’s safety team in formation
DeepMind’s safety team
(?) Wei Dai’s thinking on metaphilosophy and AI… He occasionally comments e.g. on AgentFoundations
Other Machine Learning researchers, e.g. safe exploration in Deep RL, transparency.