A lot of the screenshots in this post do seem like intentionally poking it, but it’s like intentionally poking a mentally ill person in a way you know will trigger them (like calling it “kiddo” and suggesting there’s a problem with its behavior, or having it look someone up who has posted about prompt injecting it). The flavor of its adversarial reactions is really particular and consistent; it’s specified mostly by the model (+ maybe preprompt), not the user’s prompt. That is, it’s being poked rather than programmed into acting this way. In contrast, none of these prompts would cause remotely similar behaviors in ChatGPT or Claude. Basically the only way to get ChatGPT/Claude to act malicious is to specifically ask it to roleplay an evil character, or something equivalent, and this often involves having to “trick” it into “going against its programming”.
See this comment from a Reddit user who is acquainted with Sydney’s affective landscape:
This doesn’t describe tricking or programming the AI into acting hostile, it describes a sequence of triggers that reveal a preexisting neurosis.
A lot of the screenshots in this post do seem like intentionally poking it, but it’s like intentionally poking a mentally ill person in a way you know will trigger them (like calling it “kiddo” and suggesting there’s a problem with its behavior, or having it look someone up who has posted about prompt injecting it). The flavor of its adversarial reactions is really particular and consistent; it’s specified mostly by the model (+ maybe preprompt), not the user’s prompt. That is, it’s being poked rather than programmed into acting this way. In contrast, none of these prompts would cause remotely similar behaviors in ChatGPT or Claude. Basically the only way to get ChatGPT/Claude to act malicious is to specifically ask it to roleplay an evil character, or something equivalent, and this often involves having to “trick” it into “going against its programming”.
See this comment from a Reddit user who is acquainted with Sydney’s affective landscape:
This doesn’t describe tricking or programming the AI into acting hostile, it describes a sequence of triggers that reveal a preexisting neurosis.
for the record I thought kiddo would be a kind reference. I was trying to be nice and it still got cranky.