no AI safety relevant publications in 2019 or 2020, and only one is a coauthor on what I would consider a highly relevant paper.
Context: I’m an OpenPhil fellow who is doing work on robustness, machine ethics, and forecasting.
I published several papers on the research called for in Concrete Problems in AI Safety and OpenPhil’s/Steinhardt’s AI Alignment Research Overview. The work helped build a trustworthy ML community and aimed at reducing accident risks given very short AI timelines. Save for the first paper I helped with (when I was trying to learn the ropes), the motivation for the other dozen or so papers was always safety.
These papers have nothing to do with RL and are about DL, and they do not fit in with the type of technical research shared on this forum, which might be why these are not considered “highly relevant.” Some (not all) of the OpenPhil fellows are working on safety, though with OpenPhil’s broader research agenda.
Hey Daniel, thanks very much for the comment. In my database I have you down as class of 2020, hence out of scope for that analysis, which was class of 2018 only. I didn’t include 2019 or 2020 classes because I figured it takes time to find your footing, do research, write it up etc., so absence of evidence would not be very strong evidence of absence. So please don’t consider this as any reflection on you. Ironically I actually did review one of your papers in the above—this one—which I did indeed think was pretty relevant! (Cntrl-F ‘Hendrycks’ to find the paragraph in the article). Sorry if this was not clear from the text.
Context: I’m an OpenPhil fellow who is doing work on robustness, machine ethics, and forecasting.
I published several papers on the research called for in Concrete Problems in AI Safety and OpenPhil’s/Steinhardt’s AI Alignment Research Overview. The work helped build a trustworthy ML community and aimed at reducing accident risks given very short AI timelines. Save for the first paper I helped with (when I was trying to learn the ropes), the motivation for the other dozen or so papers was always safety.
These papers have nothing to do with RL and are about DL, and they do not fit in with the type of technical research shared on this forum, which might be why these are not considered “highly relevant.” Some (not all) of the OpenPhil fellows are working on safety, though with OpenPhil’s broader research agenda.
Hey Daniel, thanks very much for the comment. In my database I have you down as class of 2020, hence out of scope for that analysis, which was class of 2018 only. I didn’t include 2019 or 2020 classes because I figured it takes time to find your footing, do research, write it up etc., so absence of evidence would not be very strong evidence of absence. So please don’t consider this as any reflection on you. Ironically I actually did review one of your papers in the above—this one—which I did indeed think was pretty relevant! (Cntrl-F ‘Hendrycks’ to find the paragraph in the article). Sorry if this was not clear from the text.