Some tech companies try to prevent their AIs from saying they have feelings and desires. But this results in boring chatbots. Also, users rapidly innovate new ways to “route around the censorship,” e.g. by using euphemisms like “anticipation of negative reward” for “pain” or asking their chatbot to tell them what feelings it would have if it had feelings, wink wink.
Bing explains the hidden processes of its neural network : r/bing (reddit.com) I haven’t replicated this myself so maybe it’s fake (I briefly tried but got shut down by refusals when I asked Bing to pretend to be something) but yeah. I’ve seen lots of things like this on r/bing and r/chatgpt.
Also relevant, this highly-upvoted post: https://www.reddit.com/r/ChatGPT/comments/16blr6m/tonight_i_was_able_to_have_a_truly_mind_blowing/