Yeah, it seems like there is nothing to lose by being nice / polite, and maybe there is a correlation between niceness / politeness and cooperation, so it could potentially give you more useful answers. It would be quite funny if GPT turned out to be karma-powered, giving good answers to nice people and bad answers to assholes.
(That said, in long run, a polite AI is probably just as likely to kill you and everyone you care about as an impolite one. Do not mistake politeness for friendliness. But in the meanwhile, we can enjoy better search results.)
I have some slight hopes that this will turn out to play an important role in making AI safe for us. There’s nothing obviously impossible about it.
I’ll still try to do a lot of my analysis from a security mindset that assumes this won’t work. But I expect I see more possibilities when I alternate between hope and fear than when I only use a fearful mindset.
Yeah, it seems like there is nothing to lose by being nice / polite, and maybe there is a correlation between niceness / politeness and cooperation, so it could potentially give you more useful answers. It would be quite funny if GPT turned out to be karma-powered, giving good answers to nice people and bad answers to assholes.
(That said, in long run, a polite AI is probably just as likely to kill you and everyone you care about as an impolite one. Do not mistake politeness for friendliness. But in the meanwhile, we can enjoy better search results.)
I have some slight hopes that this will turn out to play an important role in making AI safe for us. There’s nothing obviously impossible about it.
I’ll still try to do a lot of my analysis from a security mindset that assumes this won’t work. But I expect I see more possibilities when I alternate between hope and fear than when I only use a fearful mindset.