What can I read/look at to skill up with “alignment.”
A good place to start is the “AGI Safety Fundamentals” course reading list, which includes materials from a diverse set of AI safety research agendas. Reading this can help you figure out who in this space is doing what, and which of that you think is useful. You can also join an official iteration of the course if you want to discuss the materials with a cohort and a facilitator (you can register interest for that here). You can also join the AI Alignment slack, to discuss these and other materials and meet others who are interested in working on AI safety.
What dark horse AI/Alignment-focused companies are out there and would be willing to hire an outsider engineer?
I’m not sure what qualifies as “dark horse”, but there are plenty of AI safety organizations interested in hiring research engineers and software engineers. For these roles, your engineering skills and safety motivation typically matter more than your experience in the community. Places off the top of my head that hire engineers for AI safety work: Redwood, Anthropic, FAR, OpenAI, DeepMind. I’m sure I’ve missed others, though, so look around! These sorts of opportunities are also usually posted on the 80k job board and in AI Alignment slack.
Thanks, that’s a super helpful reading list and a hell of a deep rabbit hole. Cheers.
I’m currently skilling up my rusty ML skills and will start looking in earnest in the next couple of months for new employment in this field. Thanks for the job board link as well.
A good place to start is the “AGI Safety Fundamentals” course reading list, which includes materials from a diverse set of AI safety research agendas. Reading this can help you figure out who in this space is doing what, and which of that you think is useful. You can also join an official iteration of the course if you want to discuss the materials with a cohort and a facilitator (you can register interest for that here). You can also join the AI Alignment slack, to discuss these and other materials and meet others who are interested in working on AI safety.
I’m not sure what qualifies as “dark horse”, but there are plenty of AI safety organizations interested in hiring research engineers and software engineers. For these roles, your engineering skills and safety motivation typically matter more than your experience in the community. Places off the top of my head that hire engineers for AI safety work: Redwood, Anthropic, FAR, OpenAI, DeepMind. I’m sure I’ve missed others, though, so look around! These sorts of opportunities are also usually posted on the 80k job board and in AI Alignment slack.
Thanks, that’s a super helpful reading list and a hell of a deep rabbit hole. Cheers.
I’m currently skilling up my rusty ML skills and will start looking in earnest in the next couple of months for new employment in this field. Thanks for the job board link as well.