The first problem I see here is that cheating at D&D is exactly what we want the AI to do.
A fair point. How about changing the reward then: don’t just avoid cheating, but be sure to tell us about any way to cheat that you discover. That way, we get the benefits without the risks.
Maybe the D&D example is unfairly biasing my reply, but giving humans wish spells without guidance is the opposite of what we want.
The first problem I see here is that cheating at D&D is exactly what we want the AI to do.
A fair point. How about changing the reward then: don’t just avoid cheating, but be sure to tell us about any way to cheat that you discover. That way, we get the benefits without the risks.
Maybe the D&D example is unfairly biasing my reply, but giving humans wish spells without guidance is the opposite of what we want.