I’m not Elizabeth or Ray, but there’s a third option which I read the comment above to mean, and which I myself find plausible.
OpenAI does have roles that are obsessively aimed at reducing catastrophic risk from egregious AI misalignment. However, without more information, an outsider should not expect that those roles actually accelerate safety more than they accelerate capabilities.
Successfully increasing safety faster than capabilities requires that person to have a number of specific skills (eg political savvy, robustness to social pressure, a higher granularity strategic/technical model than most EAs have in practice, the etc.), over and above the skills that would be required to get hired for the role.
Lacking those skills, a hire for such a role is more likely to do harm than good, not primarily because they’ll be transitioned to other tasks, but because much of the work that the typical hire for such a role would end up doing either 1) doesn’t help or 2) will end up boosting OpenAI’s general capabilities more than it helps.
Furthermore, by working at OpenAI at all, they provide some legitimacy to the org as a whole, and to the existentially dangerous work happening in other parts of it, even if their work, does 0 direct harm. Someone working in such a role has to do sufficiently beneficial on-net work to overcome this baseline effect.
I’m not Elizabeth or Ray, but there’s a third option which I read the comment above to mean, and which I myself find plausible.