Goto

Collaborating Authors

 benevolence


Will Annotators Disagree? Identifying Subjectivity in Value-Laden Arguments

Homayounirad, Amir, Liscio, Enrico, Wang, Tong, Jonker, Catholijn M., Siebert, Luciano C.

arXiv.org Artificial Intelligence

Aggregating multiple annotations into a single ground truth label may hide valuable insights into annotator disagreement, particularly in tasks where subjectivity plays a crucial role. In this work, we explore methods for identifying subjectivity in recognizing the human values that motivate arguments. We evaluate two main approaches: inferring subjectivity through value prediction vs. directly identifying subjectivity. Our experiments show that direct subjectivity identification significantly improves the model performance of flagging subjective arguments. Furthermore, combining contrastive loss with binary cross-entropy loss does not improve performance but reduces the dependency on per-label subjectivity. Our proposed methods can help identify arguments that individuals may interpret differently, fostering a more nuanced annotation process.


Scientists crack the ultimate answer to the meaning of life... and it's hidden among 38M obituaries

Daily Mail - Science & tech

Trump's Epstein crisis explodes as lewd birthday letter showing president's signature is revealed Judge's'promise' let career criminal walk free to butcher Ukrainian refugee after his MOM said he should be locked up'She was so f***ed up': Carolyn Bessette's friends tell MAUREEN CALLAHAN of her secret Daddy issue, JFK Jr's murder brag that drove her mad... and why everything we know about her is a lie The chaos behind when Meghan Markle was told not to be at Queen Elizabeth II's deathbed They were locked in a dungeon inside a house of horrors. But incredible footage shows five kids' daring acts while their parents were out... and it left neighbors speechless Turn back the clock with the K-beauty retinol cream Amazon shoppers say leaves their skin'silky smooth' - and it's now $10 Scientists crack the ultimate answer to the meaning of life... and it's hidden among 38M obituaries CBS News hires a CONSERVATIVE to police interviews after Trump and Noem'deceptive' editing fury Scientist claims life on Earth was not random... but engineered Supreme Court LIFTS restrictions on Trump's immigration raids despite claims agents targeted people by race I was 52 with a collapsed'turkey neck'. Here's how I turned back the clock 10 years Plastic surgeons weigh in on Jessica Simpson's dramatic new look at VMAs as fans declare her'unrecognizable' Billionaire turns his back on Trump as he blasts President's'risky' financial move that could cost Americans their savings Trump loses appeal and must pay $83 million to E. Jean Carroll AMANDA PLATELL: Harry is'desperate' to come back to Britain and reclaim his royal role - but this fresh snub from William makes it clear why it will never happen... and why he'll never forgive his brother Scientists crack the ultimate answer to the meaning of life... and it's hidden among 38million obituaries Scientists on a mission to uncover what constitutes a life well lived found the answer after analyzing 38 million obituaries from the US spanning 30 years. Using automated text analysis tools, the team found that the most commonly celebrated values were tradition and benevolence. Nearly 80 percent of obituaries highlighted respect for customs or religion, while 76 percent emphasized caring, reliability and trustworthiness.


Value Portrait: Assessing Language Models' Values through Psychometrically and Ecologically Valid Items

Han, Jongwook, Choi, Dongmin, Song, Woojung, Lee, Eun-Ju, Jo, Yohan

arXiv.org Artificial Intelligence

The importance of benchmarks for assessing the values of language models has been pronounced due to the growing need of more authentic, human-aligned responses. However, existing benchmarks rely on human or machine annotations that are vulnerable to value-related biases. Furthermore, the tested scenarios often diverge from real-world contexts in which models are commonly used to generate text and express values. To address these issues, we propose the Value Portrait benchmark, a reliable framework for evaluating LLMs' value orientations with two key characteristics. First, the benchmark consists of items that capture real-life user-LLM interactions, enhancing the relevance of assessment results to real-world LLM usage. Second, each item is rated by human subjects based on its similarity to their own thoughts, and correlations between these ratings and the subjects' actual value scores are derived. This psychometrically validated approach ensures that items strongly correlated with specific values serve as reliable items for assessing those values. Through evaluating 44 LLMs with our benchmark, we find that these models prioritize Benevolence, Security, and Self-Direction values while placing less emphasis on Tradition, Power, and Achievement values. Also, our analysis reveals biases in how LLMs perceive various demographic groups, deviating from real human data.


Cultural Value Alignment in Large Language Models: A Prompt-based Analysis of Schwartz Values in Gemini, ChatGPT, and DeepSeek

Segerer, Robin

arXiv.org Artificial Intelligence

This study examines cultural value alignment in large language models (LLMs) by analyzing how Gemini, ChatGPT, and DeepSeek prioritize values from Schwartz's value framework. Using the 40-item Portrait Values Questionnaire, we assessed whether DeepSeek, trained on Chinese-language data, exhibits distinct value preferences compared to Western models. Results of a Bayesian ordinal regression model show that self-transcendence values (e.g., benevolence, universalism) were highly prioritized across all models, reflecting a general LLM tendency to emphasize prosocial values. However, DeepSeek uniquely downplayed self-enhancement values (e.g., power, achievement) compared to ChatGPT and Gemini, aligning with collectivist cultural tendencies. These findings suggest that LLMs reflect culturally situated biases rather than a universal ethical framework. To address value asymmetries in LLMs, we propose multi-perspective reasoning, self-reflective feedback, and dynamic contextualization. This study contributes to discussions on AI fairness, cultural neutrality, and the need for pluralistic AI alignment frameworks that integrate diverse moral perspectives.


Towards the Ultimate Programming Language: Trust and Benevolence in the Age of Artificial Intelligence

Sawicki, Bartosz, Śmiałek, Michał, Skowron, Bartłomiej

arXiv.org Artificial Intelligence

Abstract: This article explores the evolving role of programming languages in the context of artificial intelligence. It highlights the need for programming languages to ensure human understanding while eliminating unnecessary implementation details and suggests that future programs should be designed to recognize and actively support user interests. The vision includes a three-level process: using natural language for requirements, translating it into a precise system definition language, and finally optimizing the code for performance. The concept of an "Ultimate Programming Language" is introduced, emphasizing its role in maintaining human control over machines. Trust, reliability, and benevolence are identified as key elements that will enhance cooperation between humans and AI systems.


An Empirical Exploration of Trust Dynamics in LLM Supply Chains

Balayn, Agathe, Yurrita, Mireia, Rancourt, Fanny, Casati, Fabio, Gadiraju, Ujwal

arXiv.org Artificial Intelligence

With the widespread proliferation of AI systems, trust in AI is an important and timely topic to navigate. Researchers so far have largely employed a myopic view of this relationship. In particular, a limited number of relevant trustors (e.g., end-users) and trustees (i.e., AI systems) have been considered, and empirical explorations have remained in laboratory settings, potentially overlooking factors that impact human-AI relationships in the real world. In this paper, we argue for broadening the scope of studies addressing `trust in AI' by accounting for the complex and dynamic supply chains that AI systems result from. AI supply chains entail various technical artifacts that diverse individuals, organizations, and stakeholders interact with, in a variety of ways. We present insights from an in-situ, empirical study of LLM supply chains. Our work reveals additional types of trustors and trustees and new factors impacting their trust relationships. These relationships were found to be central to the development and adoption of LLMs, but they can also be the terrain for uncalibrated trust and reliance on untrustworthy LLMs. Based on these findings, we discuss the implications for research on `trust in AI'. We highlight new research opportunities and challenges concerning the appropriate study of inter-actor relationships across the supply chain and the development of calibrated trust and meaningful reliance behaviors. We also question the meaning of building trust in the LLM supply chain.


Values That Are Explicitly Present in Fairy Tales: Comparing Samples from German, Italian and Portuguese Traditions

Diaz-Faes, Alba Morollon, Murteira, Carla Sofia Ribeiro, Ruskov, Martin

arXiv.org Artificial Intelligence

Looking at how social values are represented in fairy tales can give insights about the variations in communication of values across cultures. We study how values are communicated in fairy tales from Portugal, Italy and Germany using a technique called word embedding with a compass to quantify vocabulary differences and commonalities. We study how these three national traditions differ in their explicit references to values. To do this, we specify a list of value-charged tokens, consider their word stems and analyse the distance between these in a bespoke pre-trained Word2Vec model. We triangulate and critically discuss the validity of the resulting hypotheses emerging from this quantitative model. Our claim is that this is a reusable and reproducible method for the study of the values explicitly referenced in historical corpora. Finally, our preliminary findings hint at a shared cultural understanding and the expression of values such as Benevolence, Conformity, and Universalism across the studied cultures, suggesting the potential existence of a pan-European cultural memory.


Optimal Scene Graph Planning with Large Language Model Guidance

Dai, Zhirui, Asgharivaskasi, Arash, Duong, Thai, Lin, Shusen, Tzes, Maria-Elizabeth, Pappas, George, Atanasov, Nikolay

arXiv.org Artificial Intelligence

Recent advances in metric, semantic, and topological mapping have equipped autonomous robots with semantic concept grounding capabilities to interpret natural language tasks. This work aims to leverage these new capabilities with an efficient task planning algorithm for hierarchical metric-semantic models. We consider a scene graph representation of the environment and utilize a large language model (LLM) to convert a natural language task into a linear temporal logic (LTL) automaton. Our main contribution is to enable optimal hierarchical LTL planning with LLM guidance over scene graphs. To achieve efficiency, we construct a hierarchical planning domain that captures the attributes and connectivity of the scene graph and the task automaton, and provide semantic guidance via an LLM heuristic function. To guarantee optimality, we design an LTL heuristic function that is provably consistent and supplements the potentially inadmissible LLM guidance in multi-heuristic planning. We demonstrate efficient planning of complex natural language tasks in scene graphs of virtualized real environments.


The Social Triad model of Human-Robot Interaction

Cameron, David, Collins, Emily, de Saille, Stevienna, Law, James

arXiv.org Artificial Intelligence

Despite the increasing interest in trust in human-robot interaction (HRI), there is still relatively little exploration of trust as a social construct in HRI. We propose that integration of useful models of human-human trust from psychology, highlight a potentially overlooked aspect of trust in HRI: a robot's apparent trustworthiness may indirectly relate to the user's relationship with, and opinion of, the individual or organisation deploying the robot. Our Social Triad for HRI model (User, Robot, Deployer), identifies areas for consideration in co-creating trustworthy robotics.


What does ChatGPT return about human values? Exploring value bias in ChatGPT using a descriptive value theory

Fischer, Ronald, Luczak-Roesch, Markus, Karl, Johannes A

arXiv.org Artificial Intelligence

There has been concern about ideological basis and possible discrimination in text generated by Large Language Models (LLMs). We test possible value biases in ChatGPT using a psychological value theory. We designed a simple experiment in which we used a number of different probes derived from the Schwartz basic value theory (items from the revised Portrait Value Questionnaire, the value type definitions, value names). We prompted ChatGPT via the OpenAI API repeatedly to generate text and then analyzed the generated corpus for value content with a theory-driven value dictionary using a bag of words approach. Overall, we found little evidence of explicit value bias. The results showed sufficient construct and discriminant validity for the generated text in line with the theoretical predictions of the psychological model, which suggests that the value content was carried through into the outputs with high fidelity. We saw some merging of socially oriented values, which may suggest that these values are less clearly differentiated at a linguistic level or alternatively, this mixing may reflect underlying universal human motivations. We outline some possible applications of our findings for both applications of ChatGPT for corporate usage and policy making as well as future research avenues. We also highlight possible implications of this relatively high-fidelity replication of motivational content using a linguistic model for the theorizing about human values.