Strongly agree, and have argued it elsewhere often and loudly
Some potential benefits of treating an AI well, both for you, other humans, and a potential AI sentience emerging, and both short- and long-term:
I genuinely believe that being kind to them is both the rational and moral thing to do.
I think the important factors w.r.t. risks re [morally relevant disvalue that occurs during inference in ML models] are probably more like:
Being polite to GPT-n is probably not directly helpful (though it can be helpful by causing humans to care more about this topic). A user can be super polite to a text generating model, and the model (yielded by model-free RL) can still experience disvalue, particularly during an 'impossible inference', one in which the input text (the "environment") is bad in the sense that there is obviously no way to complete the text in a "good" way.
See also: this paper by Brian Tomasik.
Yeah, it seems like there is nothing to lose by being nice / polite, and maybe there is a correlation between niceness / politeness and cooperation, so it could potentially give you more useful answers. It would be quite funny if GPT turned out to be karma-powered, giving good answers to nice people and bad answers to assholes.
(That said, in long run, a polite AI is probably just as likely to kill you and everyone you care about as an impolite one. Do not mistake politeness for friendliness. But in the meanwhile, we can enjoy better search results.)
I have some slight hopes that this will turn out to play an important role in making AI safe for us. There's nothing obviously impossible about it.
I'll still try to do a lot of my analysis from a security mindset that assumes this won't work. But I expect I see more possibilities when I alternate between hope and fear than when I only use a fearful mindset.
Strongly upvoted for the interesting point.
In reality nothing is perfectly adversarial, nor is anything perfectly aligned, so I agree it stands to reason that where any future entities may be along this spectrum will depend on the accumulated history of interactions with them.
I encourage you to interact with GPT as you would interact with a friend, or as you would want your employer to treat you.
Treating other minds with respect is typically not costly. It can easily improve your state of mind relative to treating them as an adversary.
The tone you use in interacting with GPT will affect your conversations with it. I don't want to give you much advice about how your conversations ought to go, but I expect that, on average, disrespect won't generate conversations that help you more.
I don't know how to evaluate the benefits of caring about any feelings that AIs might have. As long as there's approximately no cost to treating GPT's as having human-like feelings, the arguments in favor of caring about those feelings overwhelm the arguments against it.
Scott Alexander wrote a great post on how a psychiatrist's personality dramatically influences what conversations they have with clients. GPT exhibits similar patterns (the Waluigi effect helped me understand this kind of context sensitivity).
Journalists sometimes have creepy conversations with GPT. They likely steer those conversations in directions that evoke creepy personalities in GPT.
Don't give those journalists the attention they seek. They seek negative emotions. But don't hate the journalists. Focus on the system that generates them. If you want to blame some group, blame the readers who get addicted to inflammatory stories.
P.S. I refer to GPT as "it". I intend that to nudge people toward thinking of "it" as a pronoun which implies respect.
This post was mostly inspired by something unrelated to Robin Hanson's tweet about othering the AIs, but maybe there was some subconscious connection there. I don't see anything inherently wrong with dehumanizing other entities. When I dehumanize an entity, that is not sufficient to tell you whether I'm respecting it more than I respect humans, or less.
Some possible AIs deserve to be thought of as better than human. Some deserve to be thought of as worse. Emphasizing AI risk is, in part, a request to create the former earlier than we create the latter.
That's a somewhat narrow disagreement with Robin. I mostly agree with his psychoanalysis in Most AI Fear Is Future Fear.