We’re talking about giving the models subjective experience, not just “emotions”. You want the AI to create conscious minds inside itself and torture them to find out whether torture is bad? And then again every time it makes a decision where torture is a conceivable outcome? I’d hope we can give the AI a model that accurately predicts how humans react to stimuli without creating a conscious observer. Humans seem to be able to do that, at least..
Beware of anthropomorphizing AIs. A Really Powerful Optimization Process shouldn’t need to “suffer” for us to tell it what suffering is, and that we would like less of it.
We’re talking about giving the models subjective experience, not just “emotions”. You want the AI to create conscious minds inside itself and torture them to find out whether torture is bad? And then again every time it makes a decision where torture is a conceivable outcome? I’d hope we can give the AI a model that accurately predicts how humans react to stimuli without creating a conscious observer. Humans seem to be able to do that, at least..
Beware of anthropomorphizing AIs. A Really Powerful Optimization Process shouldn’t need to “suffer” for us to tell it what suffering is, and that we would like less of it.