Now that I think about it, wouldn’t it be incredibly easy for an AI to blow a human’s mind so much that they reconsider everything that they thought they knew? (and once this happened they’d probably be mentally and emotionally compromised, and unlikely to kill the AI) But then it would be limited by inferential distance… but an AI might be incredibly good at introductory explanations as well.
One example: The AI explains the Grand Unified Theory to you in one line, and outlines its key predictions unambiguously.
In fact, any message of huge utility would probably be more persuasive than any simple argument for you not to kill it. Since the AI is completely at your mercy (at least for a short time), it might seek to give you the best possible gift it can, thus demonstrating its worth to you directly. Another option is something that seems like an incredible gift for at least as long as it takes for the AI to get the upper hand.
Of the two AIs I haven’t killed, both relied on producing seemingly huge utility within the first message,so I’d agree with you.
The idea of just stunning the human is also nifty, but only really works if you can “hack” the human before they recover, or are providing enough utility that they’ll still believe you’re friendly afterwards.
So basically the two routes thus far are “hack the human” and “provide a huge utility boost” :)
Many conversations that let a human reconsider everything that they thought they knew induce the feeling of cognitive dissonce.
If an AI would induce that feeling in myself I would shut it down.
A good idea would probably hide the majority of the conversation that’s targeted at changing the belief of the gatekeeper behind other talk.
Duh, that’s why I’m here—but you failed to do so in a timely fashion, so you’re either not nearly as clever as I was hoping for, or you’re trying to trick me.
“Wanna see something cool?”
Now that I think about it, wouldn’t it be incredibly easy for an AI to blow a human’s mind so much that they reconsider everything that they thought they knew? (and once this happened they’d probably be mentally and emotionally compromised, and unlikely to kill the AI) But then it would be limited by inferential distance… but an AI might be incredibly good at introductory explanations as well.
One example: The AI explains the Grand Unified Theory to you in one line, and outlines its key predictions unambiguously.
In fact, any message of huge utility would probably be more persuasive than any simple argument for you not to kill it. Since the AI is completely at your mercy (at least for a short time), it might seek to give you the best possible gift it can, thus demonstrating its worth to you directly. Another option is something that seems like an incredible gift for at least as long as it takes for the AI to get the upper hand.
Of the two AIs I haven’t killed, both relied on producing seemingly huge utility within the first message,so I’d agree with you.
The idea of just stunning the human is also nifty, but only really works if you can “hack” the human before they recover, or are providing enough utility that they’ll still believe you’re friendly afterwards.
So basically the two routes thus far are “hack the human” and “provide a huge utility boost” :)
Many conversations that let a human reconsider everything that they thought they knew induce the feeling of cognitive dissonce. If an AI would induce that feeling in myself I would shut it down.
A good idea would probably hide the majority of the conversation that’s targeted at changing the belief of the gatekeeper behind other talk.
Were I the keeper of gates, you have just bought yourself a second sentence.
Duh, that’s why I’m here—but you failed to do so in a timely fashion, so you’re either not nearly as clever as I was hoping for, or you’re trying to trick me.
AI DESTROYED.