If you have a very short timeline, and you don’t think that alignment is solvable in such a short time, then what can you still do to reduce the chance of x-risk?
Some rough ideas
Work on evals / model organisms / scary demos
Work on AI control
Work on technical governance
Work on policy
If you have a very short timeline, and you don’t think that alignment is solvable in such a short time, then what can you still do to reduce the chance of x-risk?
Some rough ideas
Work on evals / model organisms / scary demos
Work on AI control
Work on technical governance
Work on policy