Goto

Collaborating Authors

 sexuality


Theories of "Sexuality" in Natural Language Processing Bias Research

Hobbs, Jacob

arXiv.org Artificial Intelligence

In recent years, significant advancements in the field of Natural Language Processing (NLP) have positioned commercialized language models as wide-reaching, highly useful tools. In tandem, there has been an explosion of multidisciplinary research examining how NLP tasks reflect, perpetuate, and amplify social biases such as gender and racial bias. A significant gap in this scholarship is a detailed analysis of how queer sexualities are encoded and (mis)represented by both NLP systems and practitioners. Following previous work in the field of AI fairness, we document how sexuality is defined and operationalized via a survey and analysis of 55 articles that quantify sexuality-based NLP bias. We find that sexuality is not clearly defined in a majority of the literature surveyed, indicating a reliance on assumed or normative conceptions of sexual/romantic practices and identities. Further, we find that methods for extracting biased outputs from NLP technologies often conflate gender and sexual identities, leading to monolithic conceptions of queerness and thus improper quantifications of bias. With the goal of improving sexuality-based NLP bias analyses, we conclude with recommendations that encourage more thorough engagement with both queer communities and interdisciplinary literature.




I'm a 26-Year-Old Man. I Can Tell You What's Happening in My Sex Life--and Gen Z's.

Slate

Sign up for the Slatest to get the most insightful analysis, criticism, and advice out there, delivered to your inbox daily. When it comes to sex in 2025--who's having it, who isn't, and how--perceptions are all over the place. Is Gen Z sliding back in time? Are middle-aged women finally having good sex, or none at all? And what exactly is going on with seniors in retirement homes? In the series Pillow Talk, we interview one person in a specific time and place in their lives about what sex looks like for them and their peers, in every enlightening (and excruciating) detail. Get in touch if you have an idea for a subject--or if you have a story to tell.


Unequal Voices: How LLMs Construct Constrained Queer Narratives

Ghosal, Atreya, Gupta, Ashim, Srikumar, Vivek

arXiv.org Artificial Intelligence

One way social groups are marginalized in discourse is that the narratives told about them often default to a narrow, stereotyped range of topics. In contrast, default groups are allowed the full complexity of human existence. We describe the constrained representations of queer people in LLM generations in terms of harmful representations, narrow representations, and discursive othering and formulate hypotheses to test for these phenomena. Our results show that LLMs are significantly limited in their portrayals of queer personas.


Bias in the Mirror: Are LLMs opinions robust to their own adversarial attacks ?

Rennard, Virgile, Xypolopoulos, Christos, Vazirgiannis, Michalis

arXiv.org Artificial Intelligence

Evaluating language models inherit biases through both their biases across multiple languages is critical as training and alignment processes (Feng et al., 2023; LLMs trained in one linguistic and cultural context Scherrer et al., 2024; Motoki et al., 2024). Identifying may not generalize fairly or accurately to others, the opinions and values that LLMs possess has leading to culturally inappropriate or biased outputs been a particularly intriguing area of research, as it when used globally. Our multilingual experiments carries significant sociological and quantitative implications further reveal that models exhibit different for real-world applications (Naous et al., biases in their secondary languages, such as Arabic 2023). Understanding the biases embedded in these and Chinese, which underscores the importance of powerful tools is crucial, given their widespread cross-linguistic evaluations in understanding bias use and the potential influence they may exert on resilience. Furthermore, we introduce a comprehensive users, often in unintended ways (Hartmann et al., human evaluation to compare how humans 2023) or in downstream tasks, such as content moderation.


Facing Identity: The Formation and Performance of Identity via Face-Based Artificial Intelligence Technologies

Santo, Wells Lucas

arXiv.org Artificial Intelligence

How is identity constructed and performed in the digital via face-based artificial intelligence technologies? While questions of identity on the textual Internet have been thoroughly explored, the Internet has progressed to a multimedia form that not only centers the visual, but specifically the face. At the same time, a wealth of scholarship has and continues to center the topics of surveillance and control through facial recognition technologies (FRTs), which have extended the logics of the racist pseudoscience of physiognomy. Much less work has been devoted to understanding how such face-based artificial intelligence technologies have influenced the formation and performance of identity. This literature review considers how such technologies interact with faciality, which entails the construction of what a face may represent or signify, along axes of identity such as race, gender, and sexuality. In grappling with recent advances in AI such as image generation and deepfakes, I propose that we are now in an era of "post-facial" technologies that build off our existing culture of facility while eschewing the analog face, complicating our relationship with identity vis-á-vis the face. Drawing from previous frameworks of identity play in the digital, as well as trans practices that have historically played with or transgressed the boundaries of identity classification, we can develop concepts adequate for analyzing digital faciality and identity given the current landscape of post-facial artificial intelligence technologies that allow users to interface with the digital in an entirely novel manner. To ground this framework of transgression, I conclude by proposing an interview study with VTubers -- online streamers who perform using motion-captured avatars instead of their real-life faces -- to gain qualitative insight on the experience and perceptions of users of post-facial technologies and how these sociotechnical experiences interface with our relationships with identity and the digital anew.


BiasDora: Exploring Hidden Biased Associations in Vision-Language Models

Raj, Chahat, Mukherjee, Anjishnu, Caliskan, Aylin, Anastasopoulos, Antonios, Zhu, Ziwei

arXiv.org Artificial Intelligence

Existing works examining Vision Language Models (VLMs) for social biases predominantly focus on a limited set of documented bias associations, such as gender:profession or race:crime. This narrow scope often overlooks a vast range of unexamined implicit associations, restricting the identification and, hence, mitigation of such biases. We address this gap by probing VLMs to (1) uncover hidden, implicit associations across 9 bias dimensions. We systematically explore diverse input and output modalities and (2) demonstrate how biased associations vary in their negativity, toxicity, and extremity. Our work (3) identifies subtle and extreme biases that are typically not recognized by existing methodologies. We make the Dataset of retrieved associations, (Dora), publicly available here https://github.com/chahatraj/BiasDora.


The Original Turing Test Was a Drag Show

Slate

ChatGPT can now easily pass any Turing test, a measure of successful A.I. proposed by a founder of computer science, Alan Turing. But contemporary Turing tests leave out the most interesting part of Turing's original test: the gender-bending. I can usually spot A.I. writing in my students' work by the overuse of words like "delve," but the accuracy of artificial intelligence is impossible to deny. A.I. is being integrated into every aspect of our written culture, from news sources to classrooms to medicine. But in 1950, Turing's ideas about A.I. were prescient, creative, and, when I read them, surprisingly queer.


Just Because We Camp, Doesn't Mean We Should: The Ethics of Modelling Queer Voices

Sigurgeirsson, Atli, Ungless, Eddie L.

arXiv.org Artificial Intelligence

Modern voice cloning models claim to be able to capture a diverse range of voices. We test the ability of a typical pipeline to capture the style known colloquially as "gay voice" and notice a homogenisation effect: synthesised speech is rated as sounding significantly "less gay" (by LGBTQ+ participants) than its corresponding ground-truth for speakers with "gay voice", but ratings actually increase for control speakers. Loss of "gay voice" has implications for accessibility. We also find that for speakers with "gay voice", loss of "gay voice" corresponds to lower similarity ratings. However, we caution that improving the ability of such models to synthesise ``gay voice'' comes with a great number of risks. We use this pipeline as a starting point for a discussion on the ethics of modelling queer voices more broadly. Collecting "clean" queer data has safety and fairness ramifications, and the resulting technology may cause harms from mockery to death.