I think this is a tricky tradeoff. There’s effectively a race between alignment and capabilities research. Better theories of how AGI is likely to be constructed will help both efforts. Which one it will help more is tough to guess.
The one thought I’d like to add is that the AI safety community may think more broadly and creatively about approaches to building AI. So I wouldn’t assume that all of this thinking has already been done.
I don’t have an answer on this, and I’ve thought about it a lot since I’ve been keeping some potential infohazard ideas under my hat for maybe the last ten years.
I think this is a tricky tradeoff. There’s effectively a race between alignment and capabilities research. Better theories of how AGI is likely to be constructed will help both efforts. Which one it will help more is tough to guess.
The one thought I’d like to add is that the AI safety community may think more broadly and creatively about approaches to building AI. So I wouldn’t assume that all of this thinking has already been done.
I don’t have an answer on this, and I’ve thought about it a lot since I’ve been keeping some potential infohazard ideas under my hat for maybe the last ten years.