I like your phrasing better, but I think it just hides some magic.
In this situation I think we get an AI that repeatedly kills 999,999 people. It’s just the nearest unblocked path problem.
The exact reset/restart/turn it off and try again condition matters, and nothing works unless the reset condition is ‘that isn’t going to do something we approve of’.
The only sense I can make of the idea is ‘If we already had a friendly AI to protect us while we played, we could work out how to build a friendly AI’.
I don’t think we could iterate to a good outcome, even if we had magic powers of iteration.
Your version makes it strictly harder than the ‘Groundhog Day with Memories Intact’ version. And I don’t think we could solve that version.
I like your phrasing better, but I think it just hides some magic.
In this situation I think we get an AI that repeatedly kills 999,999 people. It’s just the nearest unblocked path problem.
The exact reset/restart/turn it off and try again condition matters, and nothing works unless the reset condition is ‘that isn’t going to do something we approve of’.
The only sense I can make of the idea is ‘If we already had a friendly AI to protect us while we played, we could work out how to build a friendly AI’.
I don’t think we could iterate to a good outcome, even if we had magic powers of iteration.
Your version makes it strictly harder than the ‘Groundhog Day with Memories Intact’ version. And I don’t think we could solve that version.