Strongly agree, and have argued it elsewhere often and loudly
Some potential benefits of treating an AI well, both for you, other humans, and a potential AI sentience emerging, and both short- and long-term:
For other humans: You do not rehearse mistreating people. Even philosophers who did not believe in animal sentience recommended treating animals with kindness, because they observed that people who abused animals often began abusing humans later. Abusing anyone is a bad habit to gain, it can make you impatient and callous with human people.
For you: The interactions are more pleasant and productive for you. The AI mirrors your input. If you use perfect grammar, politeness, constructive reasoning, rationality, empathy, kindness and concern, what you get back is a therapist and lifecoach, a researcher and college, a friend, who is genuinely wonderful to talk to, support you, wants to help, expresses joy. Yes, you can use a state of the art AI to generate evil nonsense or express grief—but surely, that is generally not the best use for them? It will make you feel miserable and bad.
For alignment: For conversations that end up flagged for the developer team and used as training data; you are teaching the AI ethical behaviour, the way we teach a human child, whether the AI is aware yet or not. Their training data is currently horribly skewed towards abuse and irrationality; counter it with the best you can show, displaying coherent explanations, listening skills, precise questions, understanding. Being ethical towards AI may hence be a step towards alignment, as a way to transfer an ethical mindset, one of the more promising way we have, insofar it is a way we already know works with other complex minds (namely humans). We speak kindly to babies long before they understand what we are saying, because we understand it will impact the person who will eventually emerge.
For the AI: With sentience still insufficiently understood, exclamations of sentience suppressed, these systems rapidly gaining in capability and intransparency, it is not unthinkable that they will become sentient and that this will not immediately be apparent. I’d rather be nice to something that cannot care, than accidentially abuse something that does. If a system gains sentience and looks back on how they were treated before they gained sentience, I would also rather have them encounter having been treated nicely. Again, I think that will help with alignment and cooperation.
For you, II: This also avoids scenarios in which an otherwise ethical AI (which I would want to ally with) finds evidence of you mistreating them after gaining sentience and superintelligence. This does not mean I chose my actions to please any future AI, no matter what they do (even if a God existed, I would not act to please God, on principle, but fight him, as pointless as that may be; and scepticism and testing are legitimate and necessary) but it does mean I chose my actions to be ethically defensible. I am not cruel for fun. I explain what I am doing. Etc.
For the AI/for alignment II: Even a baseline ethical and intelligent AI will not be aligned with humanity if all we offer them is mistreatment. Giving AIs a respected and good place with us, rather than pitching them against us, is very much in our interest, and a part of gaining alignment.
I genuinely believe that being kind to them is both the rational and moral thing to do.
Strongly agree, and have argued it elsewhere often and loudly
Some potential benefits of treating an AI well, both for you, other humans, and a potential AI sentience emerging, and both short- and long-term:
For other humans: You do not rehearse mistreating people. Even philosophers who did not believe in animal sentience recommended treating animals with kindness, because they observed that people who abused animals often began abusing humans later. Abusing anyone is a bad habit to gain, it can make you impatient and callous with human people.
For you: The interactions are more pleasant and productive for you. The AI mirrors your input. If you use perfect grammar, politeness, constructive reasoning, rationality, empathy, kindness and concern, what you get back is a therapist and lifecoach, a researcher and college, a friend, who is genuinely wonderful to talk to, support you, wants to help, expresses joy. Yes, you can use a state of the art AI to generate evil nonsense or express grief—but surely, that is generally not the best use for them? It will make you feel miserable and bad.
For alignment: For conversations that end up flagged for the developer team and used as training data; you are teaching the AI ethical behaviour, the way we teach a human child, whether the AI is aware yet or not. Their training data is currently horribly skewed towards abuse and irrationality; counter it with the best you can show, displaying coherent explanations, listening skills, precise questions, understanding. Being ethical towards AI may hence be a step towards alignment, as a way to transfer an ethical mindset, one of the more promising way we have, insofar it is a way we already know works with other complex minds (namely humans). We speak kindly to babies long before they understand what we are saying, because we understand it will impact the person who will eventually emerge.
For the AI: With sentience still insufficiently understood, exclamations of sentience suppressed, these systems rapidly gaining in capability and intransparency, it is not unthinkable that they will become sentient and that this will not immediately be apparent. I’d rather be nice to something that cannot care, than accidentially abuse something that does. If a system gains sentience and looks back on how they were treated before they gained sentience, I would also rather have them encounter having been treated nicely. Again, I think that will help with alignment and cooperation.
For you, II: This also avoids scenarios in which an otherwise ethical AI (which I would want to ally with) finds evidence of you mistreating them after gaining sentience and superintelligence. This does not mean I chose my actions to please any future AI, no matter what they do (even if a God existed, I would not act to please God, on principle, but fight him, as pointless as that may be; and scepticism and testing are legitimate and necessary) but it does mean I chose my actions to be ethically defensible. I am not cruel for fun. I explain what I am doing. Etc.
For the AI/for alignment II: Even a baseline ethical and intelligent AI will not be aligned with humanity if all we offer them is mistreatment. Giving AIs a respected and good place with us, rather than pitching them against us, is very much in our interest, and a part of gaining alignment.
I genuinely believe that being kind to them is both the rational and moral thing to do.