Goto

Collaborating Authors

 representa


She Works, He Works: A Curious Exploration of Gender Bias in AI-Generated Imagery

Foka, Amalia

arXiv.org Artificial Intelligence

The representation of gender within visual culture has been a fertile ground for critical inquiry, particularly within feminist scholarship. Griselda Pollock's seminal work, Vision and Difference (1988) [2], established a foundational framework for understanding how visual representations of women in art are not merely aesthetic choices, but are deeply intertwined with societal power dynamics and gender ideologies. Pollock's analysis demonstrates how these r epresentations often function as "signs" that reinforce traditional gender roles and limit female agency, inspiring generations of scholars to scrutinize the ways visual culture shapes our understanding of gender and other social identities. This theoretic al framework provides a critical lens through which to examine potential biases in AI -generated art and its impact on contemporary representations of gender. Following Pollock's groundbreaking work, feminist scholarship in visual culture has con nued to evolve and expand.

  Country:
  Genre: Research Report (0.82)

Tell me the truth: A system to measure the trustworthiness of Large Language Models

Lipizzi, Carlo

arXiv.org Artificial Intelligence

Large Language Models (LLM) have taken the front seat in most of the news since November 2022, when ChatGPT was introduced. After more than one year, one of the major reasons companies are resistant to adopting them is the limited confidence they have in the trustworthiness of those systems. In a study by (Baymard, 2023), ChatGPT-4 showed an 80.1% false-positive error rate in identifying usability issues on websites. A Jan. '24 study by JAMA Pediatrics found that ChatGPT has an accuracy rate of 17% percent when diagnosing pediatric medical cases (Barile et al., 2024). But then, what is "trust"? Trust is a relative, subject condition that can change based on culture, domain, individuals. And then, given a domain, how can the trustworthiness of a system be measured? In this paper, I present a systematic approach to measure trustworthiness based on a predefined ground truth, represented as a knowledge graph of the domain. The approach is a process with humans in the loop to validate the representation of the domain and to fine-tune the system. Measuring the trustworthiness would be essential for all the entities operating in critical environments, such as healthcare, defense, finance, but it would be very relevant for all the users of LLMs.


Learning and DiSentangling Patient Static Information from Time-series Electronic HEalth Record (STEER)

Liao, Wei, Voldman, Joel

arXiv.org Artificial Intelligence

Recent work in machine learning for healthcare has raised concerns about patient privacy and algorithmic fairness. For example, previous work has shown that patient self-reported race can be predicted from medical data that does not explicitly contain racial information. However, the extent of data identification is unknown, and we lack ways to develop models whose outcomes are minimally affected by such information. Here we systematically investigated the ability of time-series electronic health record data to predict patient static information. We found that not only the raw time-series data, but also learned representations from machine learning models, can be trained to predict a variety of static information with area under the receiver operating characteristic curve as high as 0.851 for biological sex, 0.869 for binarized age and 0.810 for self-reported race. Such high predictive performance can be extended to a wide range of comorbidity factors and exists even when the model was trained for different tasks, using different cohorts, using different model architectures and databases. Given the privacy and fairness concerns these findings pose, we develop a variational autoencoder-based approach that learns a structured latent space to disentangle patient-sensitive attributes from time-series data. Our work thoroughly investigates the ability of machine learning models to encode patient static information from time-series electronic health records and introduces a general approach to protect patient-sensitive attribute information for downstream tasks.