Any other ideas for “decoupled” AIs, or risks that apply to this approach in general?
If the question is about all the risks that apply, rather than special risks with this specific approach, then I’ll note that the usual risks from the inner alignment problem seem to apply.
If the question is about all the risks that apply, rather than special risks with this specific approach, then I’ll note that the usual risks from the inner alignment problem seem to apply.