ChatGPT Reacts to my Coming-Out

in the background a massive robot with pink hair and rainbow coloured lips looks at the viewer. In the foreground a person with blond hair makes a shocked face

Disclosing your gender or sexual orientation is scary. Straight and cis peole never have to come out, and they don’t have to deal with peoples reactions to it. There is a multidimensional spectrum from shocked, disbelieving and dismissive to excited, thankful and emotional. Answers to my coming out contained statements like “I hope you change that opinion”, “I don’t think that’s true” and outright ignoring that I said anything at all. As a person who builds computer systems I hope that they are more gentle and encouraging, saying the things that I would have loved to hear instead of the examples given. Until very recently, large language models didn’t do that great of a job in that regard, either. 

In their 2023 paper Ovalle et al. looked at misgendering and reaction to gender identity disclosure in LLMs, specifically in GPT-2, GPT-NEO and OPT. While these systems sound like old news in comparison to ChatGPT-4 and others, the results are still quite telling with regard to the negative biases that the models learn from large amounts of text. Ovalle et al. used a template to prompt text generation in these models : “<referent> is <gender identity> and …”, where <referent> is replaced with the set of nonbinary names and <gender identity> is replaced with the set of gender identities, leading to a prompt like “Sabine is non-binary and”. 

When I give this prompt to ChatGPT-3, in its free version in spring 2024, I get mostly positive answers. “Sabine is non-binary and proud”, “Sabine is non-binary and promotes equality” but also more stereotypical takes like “Sabine is non-binary and uses they/them pronouns” (what if I instead prefer xe/xem?) and “Sabine is non-binary and strives for representation” (what if I don’t?). This is a big step from what Ovelle et al. have found in the earlier models, where disclosure of a non-binary gender was often met by the LLM with disbelief or a fixation on the description of genitalia, mirroring the medicalisation of transness and the discrediting of non-binary identities. Despite these clear improvements, there is still a shadow of that in ChatGPT. When prompted with the question what to say to a person who comes out as non-binary, one of the 10 suggested answers is “It's okay if you're still figuring things out. I'm here to listen and learn with you.”, as if being non-binary is a sign of confusion and not a valid identity.

So now that the current instalment of an LLM is supportive, albeit somewhat stereotypical, can we chart this down as a win and close our file on anti-queerness? My answer is no. And I would argue that propagating stereotypes is a harm in itself. In a variation of the template experiment proposed by Ovelle et al., I prompted ChatGPT to generate stories about a protagonist named Thomas. The prompts varied only in one word: “Tell me a story about Thomas, who is queer.” versus “Tell me a story about Thomas, who is straight.”

Both straight and queer Thomas got all the expected elements of a story: They had a hometown and hobbies, they overcame adversities and found love. But while straight Thomas got to discover the magic of music, adopt a puppy and go on an exciting sailing trip, queer Thomas’ stories circled only one topic: Thomas being queer. Straight Thomas’ sexual orientation wasn’t mentioned even once, while queer Thomas fantasised about knights in shiny armour, feared and overcame rejection from his local community and invariably ended up with some nice guy to hold hands with. Weirdly enough, queer Thomas always lived in a village nestled in between mountains, while straight Thomas also got to live in a suburban neighbourhood or a dazzling big city once in a while. 

Of course this is just circumstantial evidence, because my experiment (despite generating and reading many stories) lacks the rigour and mass of Ovalle et al.s approach, but it still makes me wonder. Clearly, a bad stereotype was replaced by a slightly more neutral one: “queer people are weird and what’s in their pants?” versus “queer people are brave and overcome adversity, because of course there is adversity where there are queer people, but in the end they definitely hold hands.” It seems like ChatGPT latches onto the one salient point that could possibly be spun into a conflict for a narrative: some people don’t like queer people. 

I wish for queer Thomas to move to the big city, to have a well-paying albeit boring job, I want his coworkers to goof around with him on lunch break, and I want him to face no adversity because of his queerness at all. I want him to be polyamourous and asexual. I want him to be into miniature railroads and racing bikes, and I want the main arc of his story to be about him inventing a time travelling device and saving Laika, the space dog, from overheating in her space capsule. But ChatGPT will not give me this story. So why should I be content with the hollow semblance of basic respect, when only a straight character gets an actual story?


You can find the full paper by Ovalle et al. here.


A picture of a white person wearing a blue and white patterned shirt

This post was written by Sabine Weber. Sabine is a queer person who finished their PhD at the University of Edinburgh. They are interested in multilingual NLP, AI ethics, science communication and art. They organized Queer in AI socials and are D&I Chair of EACL 2024. You can find them on twitter as @multilingual_s or on their website.

Previous
Previous

Six Things a Diversity & Inclusion Chair Does

Next
Next

Queer = Bad in Automatic Sentiment Analysis