could probably make it avoid stabbing you to death in all these cases by giving it negative utility for your boundary being violated
yeah
but in that case it might also interfere with you getting vaccinated or getting plastic surgery
I think this can be solved with some notion of consent
Maybe ideally, though, «boundaries» don’t have to be respected, they just have to be defended by the organism inside. Maybe everyone has their own AGI to defend their «boundary».
But I strongly doubt this could be used to prevent it from making propaganda to trick you.
I think this is fundamentally different than stabbing someone. You have the power to resist propaganda. This is where I disagree with a lot of people on LW: manipulation is not intrinsically unavoidable.[1]
There’s an exception if you can literally scan someone’s brain, predict it forward in time, and reverse engineer the outputs that you want. But besides that— besides actions with very high information costs— I think it’s basically not possible (in the same way decrypting an encrypted message is theoretically possible but not practical).
Maybe ideally, though, «boundaries» don’t have to be respected, they just have to be defended by the organism inside. Maybe everyone has their own AGI to defend their «boundary».
I think this requires relatively low inequality.
I think this is fundamentally different than stabbing someone. You have the power to resist propaganda. This is where I disagree with a lot of people on LW: manipulation is not intrinsically unavoidable.
I disagree but it might not be super relevant to resolve here? Idk.
yeah
I think this can be solved with some notion of consent
Maybe ideally, though, «boundaries» don’t have to be respected, they just have to be defended by the organism inside. Maybe everyone has their own AGI to defend their «boundary».
I think this is fundamentally different than stabbing someone. You have the power to resist propaganda. This is where I disagree with a lot of people on LW: manipulation is not intrinsically unavoidable.[1]
There’s an exception if you can literally scan someone’s brain, predict it forward in time, and reverse engineer the outputs that you want. But besides that— besides actions with very high information costs— I think it’s basically not possible (in the same way decrypting an encrypted message is theoretically possible but not practical).
I think this requires relatively low inequality.
I disagree but it might not be super relevant to resolve here? Idk.
Yeah, but this has been the case for all of history and probably will forever be.
But the key point is that «boundaries» are intersubjectively definable, and it’s not just that everyone’s intrinsic self is smearing over each other.
Looking forward to your comments on the draft I sent you:)