Note that there are two parts to this, both big, hairy, and unsolved:
1) teach the AI to know what many groups of humans would consider “cheating”. I expect “cheating” is only a subset of bad behaviors, and this is just an instance of “understand human CEV”.
2) motivate the AI to not cheat. Unless cheating would help further human interest, maybe.
Yes. I am suggesting to teach AI to identify cheating as a comparatively simple way of making an AI friendly. For what other reason did you think I suggested it?
The grandparent suggests that you need a separate solution to make your solution work. The claim seems to be that you can’t solve FAI this way, because you’d need to have already solved the problem in order to make your idea stretch far enough.
Note that there are two parts to this, both big, hairy, and unsolved: 1) teach the AI to know what many groups of humans would consider “cheating”. I expect “cheating” is only a subset of bad behaviors, and this is just an instance of “understand human CEV”. 2) motivate the AI to not cheat. Unless cheating would help further human interest, maybe.
In short, “solve friendly AI”.
Yes. I am suggesting to teach AI to identify cheating as a comparatively simple way of making an AI friendly. For what other reason did you think I suggested it?
The grandparent suggests that you need a separate solution to make your solution work. The claim seems to be that you can’t solve FAI this way, because you’d need to have already solved the problem in order to make your idea stretch far enough.