That’s a defensible position, if you care about the utility of beings that don’t currently exist, to the extent that you trade the utility of currently existing beings to create new, happier ones.
The point is that the result of total utility maximization is unlikely to be something we’d recognize as people, even wireheads or Super Happy People.
Who cares about humans exactly? I care about utility.
That is nonsense. Utility is usefulness to people. If there are no humans there is no utility. An AI that could become convinced that “humans are not an efficient way to generate utility” would be what is referred to as a paperclipper.
This is why I don’t like the utile jargon. It makes it sound as though utility was something that could be measured independently of human emotions. Perhaps some kind of substance. But if statements about utility are not translated back to statements about human action or goals then they are completely meaningless.
Utility is usefulness to people. If there are no humans there is no utility.
Utility is goodness measured according to some standard of goodness; that standard doesn’t have to reference human beings. In my most optimistic visions of a far future, human values outlive the human race.
Utility is goodness measured according to some standard of goodness; that standard doesn’t have to reference human beings. In my most optimistic visions of a far future, human values outlive the human race.
Are we using the same definition of “human being”? We would not have to be biologically identical to what we are now in order to be people. But human values without humans also sounds meaningless to me. There are no values atoms or goodness atoms sitting around somewhere. To be good or to be valuable something must be good or valuable by the standards of some person. So there would have to be somebody around to do the valuing. But the standards don’t have to be explicit or objective.
Utility as I care about it is probably the result of information processing. Not clear why information should only be able to be processed in that way by human type minds, let alone fleshy ones.
Starting with the assumption of utilitarianism, I believe you’re correct. I think the folks working on this stuff assign a low probability to “kill all humans” being Friendly. But I’m pretty sure people aren’t supposed to speculate about the output of CEV.
But perhaps the proportion of FAIs that ‘kill all humans’ is large.
Maybe probability you estimate for that to happen is high, but “proportion” doesn’t makes sense, since FAI is defined as an agent acting for specific preference, so FAIs have to agree on what to do.
Who cares about humans exactly? I care about utility. If the AI thinks humans aren’t an efficient way of generating utility, we should be eliminated.
That’s a defensible position, if you care about the utility of beings that don’t currently exist, to the extent that you trade the utility of currently existing beings to create new, happier ones.
The point is that the result of total utility maximization is unlikely to be something we’d recognize as people, even wireheads or Super Happy People.
That is nonsense. Utility is usefulness to people. If there are no humans there is no utility. An AI that could become convinced that “humans are not an efficient way to generate utility” would be what is referred to as a paperclipper.
This is why I don’t like the utile jargon. It makes it sound as though utility was something that could be measured independently of human emotions. Perhaps some kind of substance. But if statements about utility are not translated back to statements about human action or goals then they are completely meaningless.
Utility is goodness measured according to some standard of goodness; that standard doesn’t have to reference human beings. In my most optimistic visions of a far future, human values outlive the human race.
Are we using the same definition of “human being”? We would not have to be biologically identical to what we are now in order to be people. But human values without humans also sounds meaningless to me. There are no values atoms or goodness atoms sitting around somewhere. To be good or to be valuable something must be good or valuable by the standards of some person. So there would have to be somebody around to do the valuing. But the standards don’t have to be explicit or objective.
Utility as I care about it is probably the result of information processing. Not clear why information should only be able to be processed in that way by human type minds, let alone fleshy ones.
Starting with the assumption of utilitarianism, I believe you’re correct. I think the folks working on this stuff assign a low probability to “kill all humans” being Friendly. But I’m pretty sure people aren’t supposed to speculate about the output of CEV.
Probably the proportion of ‘kill all humans’ AIs that are friendly is low. But perhaps the proportion of FAIs that ‘kill all humans’ is large.
That depends on your definition of Friendly, which in turn depends on your values.
Maybe probability you estimate for that to happen is high, but “proportion” doesn’t makes sense, since FAI is defined as an agent acting for specific preference, so FAIs have to agree on what to do.
OK, I’m new to this.