I don’t fully understand why you’re concerned about the possibility of misaligned AI, considering that the alignment problem has essentially been solved. We know how to ensure alignment. ChaosGPT, for example, is aligned with the values of an individual who requested it to pretend to be evil.
So we can bvring about a kind of negative alignment in systems that aren’t agentive?
So we can bvring about a kind of negative alignment in systems that aren’t agentive?