Is there much the reduced-impact agent with reward shaping could do that an agent using human mimicry couldn’t?
Perhaps it could improve over mimicry by being able to consider all actions, while a human mimic would only in effect consider the actions a human would. But I don’t think there are usually many single-step actions to choose from, so I’m guessing this isn’t a big benefit. Could the performance improvement come from better understanding the current state than mimics could? I’m not sure when this would make a big difference, though.
I’m also still concerned the reduced-impact agent would find some clever way to cause devastation while avoiding the impact penalty, but I’m less concerned about human mimics causing devastation. Are there other, major risks to using mimicry that the reduced-impact agent avoids?
Is there much the reduced-impact agent with reward shaping could do that an agent using human mimicry couldn’t?
Perhaps it could improve over mimicry by being able to consider all actions, while a human mimic would only in effect consider the actions a human would. But I don’t think there are usually many single-step actions to choose from, so I’m guessing this isn’t a big benefit. Could the performance improvement come from better understanding the current state than mimics could? I’m not sure when this would make a big difference, though.
I’m also still concerned the reduced-impact agent would find some clever way to cause devastation while avoiding the impact penalty, but I’m less concerned about human mimics causing devastation. Are there other, major risks to using mimicry that the reduced-impact agent avoids?