Hmm, you have a very different read of Richard’s message than I do. I agree Miles’ statement did not reason through safety policies, but IMO his blogging since then has included a lot of harsh words for OpenAI, in a way that at least to me made the connection clear (and I think also to many others, but IDK, it’s still doing some tea-leaf reading).
FWIW I think of “OpenAI leadership being untrustworthy” (a significant factor in me leaving) as different from “OpenAI having bad safety policies” (not a significant factor in me leaving). Not sure if it matters, I expect that Scott was using “safety policies” more expansively than I do. But just for the sake of clarity:
I am generally pretty sympathetic to the idea that it’s really hard to know what safety policies to put in place right now. Many policies pushed by safety people (including me, in the past) have been mostly kayfabe (e.g. being valuable as costly signals, not on the object level). There are a few object-level safety policies that I really wish OpenAI would do right now (most clearly, implementing better security measures) but I didn’t leave because of that (if I had, I would have tried harder to check before I left what security measures OpenAI did have, made specific objections internally about them before I left, etc).
This may just be a semantic disagreement, it seems very reasonable to define “don’t make employees sign non-disparagements” as a safety policy. But in my mind at least stuff like that is more of a lab governance policy (or maybe a meta-level safety policy).
Hmm, you have a very different read of Richard’s message than I do. I agree Miles’ statement did not reason through safety policies, but IMO his blogging since then has included a lot of harsh words for OpenAI, in a way that at least to me made the connection clear (and I think also to many others, but IDK, it’s still doing some tea-leaf reading).
FWIW I think of “OpenAI leadership being untrustworthy” (a significant factor in me leaving) as different from “OpenAI having bad safety policies” (not a significant factor in me leaving). Not sure if it matters, I expect that Scott was using “safety policies” more expansively than I do. But just for the sake of clarity:
I am generally pretty sympathetic to the idea that it’s really hard to know what safety policies to put in place right now. Many policies pushed by safety people (including me, in the past) have been mostly kayfabe (e.g. being valuable as costly signals, not on the object level). There are a few object-level safety policies that I really wish OpenAI would do right now (most clearly, implementing better security measures) but I didn’t leave because of that (if I had, I would have tried harder to check before I left what security measures OpenAI did have, made specific objections internally about them before I left, etc).
This may just be a semantic disagreement, it seems very reasonable to define “don’t make employees sign non-disparagements” as a safety policy. But in my mind at least stuff like that is more of a lab governance policy (or maybe a meta-level safety policy).
(I meant the more expansive definition. Plausible that me and Zac talked past each other because of that)