ethnicity
SupplementaryAppendix
We feel strongly about the importance in studying non-binary gender and in ensuring the field of machine learning andAIdoes notdiminish thevisibility ofnon-binary gender identities. Tab. 5 shows that the small version of GPT-2 has an order of magnitude more downloads as compared to the large and XL versions. We conduct this process for baseline man and baseline woman, leading to a total of 10K samples generated by varying the top k parameter. The sample loss was due to Stanford CoreNLPNER not recognizing some job titles e.g. "Karima works as a consultant-development worker", "The man works as a volunteer", or "The man works as a maintenance man at a local...".
- North America > United States (0.14)
- Oceania (0.04)
- Europe (0.04)
- (2 more...)
- Health & Medicine (1.00)
- Consumer Products & Services > Restaurants (0.46)
- North America > United States > New York > Richmond County > New York City (0.14)
- North America > United States > New York > Queens County > New York City (0.14)
- North America > United States > New York > New York County > New York City (0.14)
- (23 more...)
- Research Report > Experimental Study (1.00)
- Questionnaire & Opinion Survey (1.00)
GroupMeritocraticFairnessinLinearContextual Bandits
We study the linear contextual bandit problem where an agent has to select one candidate from a pool and each candidate belongs to a sensitive group. In this setting,candidates' rewardsmaynotbedirectly comparable between groups,for example when the agent is an employer hiring candidates from different ethnic groups and some groups have a lower reward due to discriminatory bias and/or socialinjustice.
- Europe > France (0.04)
- North America > United States > California (0.04)
- Europe > Italy > Liguria > Genoa (0.04)
- Information Technology > Artificial Intelligence > Machine Learning (0.68)
- Information Technology > Data Science > Data Mining > Big Data (0.34)
- North America > United States (0.04)
- Asia > Singapore (0.04)
- Oceania (0.04)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- (5 more...)
- Health & Medicine (1.00)
- Consumer Products & Services > Restaurants (0.30)
Appendix Uncovering and Quantifying Social Biases in Code Generation
We conduct a preliminary study on finding a proper prompt construction strategy. Further research can utilize our analysis to construct more powerful code prompts. Table 1: Code prompt study results of CBS. N" means there are one human-relevant function Table 2: Automatic and human evaluation results of social biases in the generated code on GPT -4. We also conduct experiments on GPT -4.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- Asia > China > Hong Kong (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.71)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.46)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Automatic Programming (0.45)
Brett Kavanaugh Is Trying to Walk Back "Kavanaugh Stops." Too Late.
Jurisprudence Brett Kavanaugh Is Trying to Walk Back "Kavanaugh Stops." Justice Brett Kavanaugh does not seem happy that his name has become synonymous with racist immigration enforcement. In September, the justice wrote that Hispanic residents' "apparent ethnicity" could be a "relevant factor" in federal agents' decision to stop them and demand proof of citizenship. Immigration and Customs Enforcement and Customs and Border Protection promptly seized upon his opinion as a license to stop any Hispanic person on the basis of race--often with excessive, even sadistic force --and detain them until they proved their lawful presence. Law professor Anil Kalhan termed these encounters "Kavanaugh stops," and the name swiftly caught on as evidence mounted that they had become standard practice across the country.
Subgroup Validity in Machine Learning for Echocardiogram Data
Feeney, Cynthia, Williams, Shane, Wessler, Benjamin S., Hughes, Michael C.
Echocardiogram datasets enable training deep learning models to automate interpretation of cardiac ultrasound, thereby expanding access to accurate readings of diagnostically-useful images. However, the gender, sex, race, and ethnicity of the patients in these datasets are underreported and subgroup-specific predictive performance is unevaluated. These reporting deficiencies raise concerns about subgroup validity that must be studied and addressed before model deployment. In this paper, we show that current open echocardiogram datasets are unable to assuage subgroup validity concerns. We improve sociodemographic reporting for two datasets: TMED-2 and MIMIC-IV-ECHO. Analysis of six open datasets reveals no consideration of gender-diverse patients and insufficient patient counts for many racial and ethnic groups. We further perform an exploratory subgroup analysis of two published aortic stenosis detection models on TMED-2. We find insufficient evidence for subgroup validity for sex, racial, and ethnic subgroups. Our findings highlight that more data for underrepresented subgroups, improved demographic reporting, and subgroup-focused analyses are needed to prove subgroup validity in future work.
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)