Deception is only a useful strategy for someone who is a) under surveillance and b) subject to constant pressures to do things differently than one would otherwise do them.
To make the AI be non-deceptive, you have three options: (a) make this fact be false; (b) make the AI fail to notice this truth; (c) prevent the AI from taking advantage of this truth.
B and C in the quote become my A and B when “this truth” in the quote are swapped for “any truth.” You can make this fact be false by not doing your B and C or my A and B.
whereas truths are all tangled together.
I think it would be worthwhile to point out that if truths are all tangled together, then your truths and its truths ought to be tangled together, too. The only situation where that wouldn’t be the case is an adversarial situation. But in your B and C cases, this is an adversarial situation, albeit one you brought upon it rather than it upon you.
But even in an adversarial situation, it should still be the case that truths are all tangled together. Therefore, there shouldn’t really be any facts you wouldn’t want it to know about—unless there is another fact that you know that causes you to not want it to discover that fact or a different fact.
If so, then what would happen if it were to discover that fact as well, in addition to the one you didn’t want it to know?
Deception is only a useful strategy for someone who is a) under surveillance and b) subject to constant pressures to do things differently than one would otherwise do them.
B and C in the quote become my A and B when “this truth” in the quote are swapped for “any truth.” You can make this fact be false by not doing your B and C or my A and B.
I think it would be worthwhile to point out that if truths are all tangled together, then your truths and its truths ought to be tangled together, too. The only situation where that wouldn’t be the case is an adversarial situation. But in your B and C cases, this is an adversarial situation, albeit one you brought upon it rather than it upon you.
But even in an adversarial situation, it should still be the case that truths are all tangled together. Therefore, there shouldn’t really be any facts you wouldn’t want it to know about—unless there is another fact that you know that causes you to not want it to discover that fact or a different fact.
If so, then what would happen if it were to discover that fact as well, in addition to the one you didn’t want it to know?