Is this the same value payload that makes activists fight over language to make human biases work for their side? I don’t think this problem translates to AI: If the AGIs find that some metric induces some bias, each can compensate for it.
Is this the same value payload that makes activists fight over language to make human biases work for their side? I don’t think this problem translates to AI: If the AGIs find that some metric induces some bias, each can compensate for it.