Well, yes, it will be shaped by what it learns, but it’s not what you train it on that matters, since you don’t get to limit the inputs and then hope it only learns “good behavior”. Human values are vague and complex, and not something to optimize for, but more of a rough guardrail. All of human output, informational and physical, is relevant here. “Safe” training data is asking for trouble once the AI learns that the real world is not at all like training data.
Well, yes, it will be shaped by what it learns, but it’s not what you train it on that matters, since you don’t get to limit the inputs and then hope it only learns “good behavior”. Human values are vague and complex, and not something to optimize for, but more of a rough guardrail. All of human output, informational and physical, is relevant here. “Safe” training data is asking for trouble once the AI learns that the real world is not at all like training data.