If an AGI reliably understands ‘shut down now’, it probably also reliably understands ‘translate this document into Russian’ but that doesn’t necessarily mean it can do anything with ‘bring about world peace’.
Unfortunately, it can, and that is one of the reasons we have to be careful. I don’t want the entire population of the planet to be forcibly sedated.
I don’t want the entire population of the planet to be forcibly sedated.
Leaving aside other reasons why that scenario is unrealistic, it does indeed illustrate why part of building a system that can reliably figure out what you mean by simple instructions, is making sure that when it’s out of its depth, it stops with an error message or request for clarification instead of guessing.
Unfortunately, it can, and that is one of the reasons we have to be careful. I don’t want the entire population of the planet to be forcibly sedated.
Leaving aside other reasons why that scenario is unrealistic, it does indeed illustrate why part of building a system that can reliably figure out what you mean by simple instructions, is making sure that when it’s out of its depth, it stops with an error message or request for clarification instead of guessing.
I think the problem is knowing when not to believe humans know what they actually want.