human agency
A First-Principles Based Risk Assessment Framework and the IEEE P3396 Standard
Tong, Richard J., Cortês, Marina, DeFalco, Jeanine A., Underwood, Mark, Zalewski, Janusz
Generative Artificial Intelligence (AI) is enabling unprecedented automation in content creation and decision support, but it also raises novel risks. This paper presents a first-principles risk assessment framework underlying the IEEE P3396 Recommended Practice for AI Risk, Safety, Trustworthiness, and Responsibility. We distinguish between process risks (risks arising from how AI systems are built or operated) and outcome risks (risks manifest in the AI system's outputs and their real-world effects), arguing that generative AI governance should prioritize outcome risks. Central to our approach is an information-centric ontology that classifies AI-generated outputs into four fundamental categories: (1) Perception-level information, (2) Knowledge-level information, (3) Decision/Action plan information, and (4) Control tokens (access or resource directives). This classification allows systematic identification of harms and more precise attribution of responsibility to stakeholders (developers, deployers, users, regulators) based on the nature of the information produced. We illustrate how each information type entails distinct outcome risks (e.g. deception, misinformation, unsafe recommendations, security breaches) and requires tailored risk metrics and mitigations. By grounding the framework in the essence of information, human agency, and cognition, we align risk evaluation with how AI outputs influence human understanding and action. The result is a principled approach to AI risk that supports clear accountability and targeted safeguards, in contrast to broad application-based risk categorizations. We include example tables mapping information types to risks and responsibilities. This work aims to inform the IEEE P3396 Recommended Practice and broader AI governance with a rigorous, first-principles foundation for assessing generative AI risks while enabling responsible innovation.
- North America > United States > Florida > Hillsborough County > University (0.04)
- North America > United States > California (0.04)
- Europe > United Kingdom > North Sea > Southern North Sea (0.04)
- (3 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.75)
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
Sturgeon, Benjamin, Samuelson, Daniel, Haimes, Jacob, Anthis, Jacy Reese
As humans delegate more tasks and decisions to artificial intelligence (AI), we risk losing control of our individual and collective futures. Relatively simple algorithmic systems already steer human decision-making, such as social media feed algorithms that lead people to unintentionally and absent-mindedly scroll through engagement-optimized content. In this paper, we develop the idea of human agency by integrating philosophical and scientific theories of agency with AI-assisted evaluation methods: using large language models (LLMs) to simulate and validate user queries and to evaluate AI responses. We develop HumanAgencyBench (HAB), a scalable and adaptive benchmark with six dimensions of human agency based on typical AI use cases. HAB measures the tendency of an AI assistant or agent to Ask Clarifying Questions, Avoid Value Manipulation, Correct Misinformation, Defer Important Decisions, Encourage Learning, and Maintain Social Boundaries. We find low-to-moderate agency support in contemporary LLM-based assistants and substantial variation across system developers and dimensions. For example, while Anthropic LLMs most support human agency overall, they are the least supportive LLMs in terms of Avoid Value Manipulation. Agency support does not appear to consistently result from increasing LLM capabilities or instruction-following behavior (e.g., RLHF), and we encourage a shift towards more robust safety and alignment targets.
- Europe > Austria > Vienna (0.14)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > Hawaii > Honolulu County > Honolulu (0.04)
- (15 more...)
- Education > Educational Setting (0.93)
- Health & Medicine > Consumer Health (0.67)
- Media > News (0.67)
- (3 more...)
Reid Hoffman: 'Start using AI deeply. It is a huge intelligence amplifier'
Reid Hoffman is a prominent Silicon Valley billionaire entrepreneur and investor known for co-founding the professional social networking site LinkedIn, now owned by Microsoft. The longtime Democrat donor threw his support behind Kamala Harris in the race for the White House. Hoffman spoke to the Observer about technology in the new political milieu and his new book about our future with artificial intelligence, Superagency. The book, while not ignoring the problems that AI might cause, argues that the technology is poised to give us cognitive superpowers that will increase our individual and collective human agency, creating a state of widespread empowerment for society. You have a vested interest in being positive about AI, including a company focused on conversational AI for business, Inflection AI.
- North America > United States > California (0.24)
- Europe > Russia (0.04)
- Asia > Russia (0.04)
- Asia > China (0.04)
- Law (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
Elon Musk's A.I.-Fuelled War on Human Agency
Not long ago, the American public could have been forgiven for thinking of Elon Musk's vaunted Department of Government Efficiency (DOGE) as a version of a familiar Republican cost-cutting, government-shrinking project. The man who took over Twitter and slashed its staff by around eighty per cent would take a similarly aggressive tack against bureaucratic inefficiency, reining in budgets and laying off federal employees. In the past couple of weeks, though, it's become clear that Musk's aim within the Trump Administration goes further: he wants not only to reduce the U.S. government but to install his own technological vision of the future at its heart. To run his agency, Musk brought on a group of tech-company managers and inexperienced twentysomethings whose credentials included internships at SpaceX. We watched as this crew began interrogating federal employees about their jobs, interfering with the system that controls payments at the Treasury Department, and trawling government budgets while Musk used X, the social platform he owns, to call out the agencies and programs in his crosshairs.
- North America > United States > California (0.06)
- North America > United States > Wisconsin (0.05)
- Information Technology > Communications > Social Media (0.88)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.75)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.51)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.51)
Explainability Paths for Sustained Artistic Practice with AI
Tecks, Austin, Peschlow, Thomas, Vigliensoni, Gabriel
The development of AI-driven generative audio mirrors broader AI trends, often prioritizing immediate accessibility at the expense of explainability. Consequently, integrating such tools into sustained artistic practice remains a significant challenge. In this paper, we explore several paths to improve explainability, drawing primarily from our research-creation practice in training and implementing generative audio models. As practical provisions for improved explainability, we highlight human agency over training materials, the viability of small-scale datasets, the facilitation of the iterative creative process, and the integration of interactive machine learning as a mapping tool. Importantly, these steps aim to enhance human agency over generative AI systems not only during model inference, but also when curating and preprocessing training data as well as during the training phase of models.
- North America > United States > Illinois > Cook County > Chicago (0.06)
- North America > United States > New York > New York County > New York City (0.05)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > Canada > Ontario > Waterloo Region > Waterloo (0.04)
Intent-aligned AI systems deplete human agency: the need for agency foundations research in AI safety
Mitelut, Catalin, Smith, Ben, Vamplew, Peter
The rapid advancement of artificial intelligence (AI) systems suggests that artificial general intelligence (AGI) systems may soon arrive. Many researchers are concerned that AIs and AGIs will harm humans via intentional misuse (AI-misuse) or through accidents (AI-accidents). In respect of AI-accidents, there is an increasing effort focused on developing algorithms and paradigms that ensure AI systems are aligned to what humans intend, e.g. AI systems that yield actions or recommendations that humans might judge as consistent with their intentions and goals. Here we argue that alignment to human intent is insufficient for safe AI systems and that preservation of long-term agency of humans may be a more robust standard, and one that needs to be separated explicitly and a priori during optimization. We argue that AI systems can reshape human intention and discuss the lack of biological and psychological mechanisms that protect humans from loss of agency. We provide the first formal definition of agency-preserving AI-human interactions which focuses on forward-looking agency evaluations and argue that AI systems - not humans - must be increasingly tasked with making these evaluations. We show how agency loss can occur in simple environments containing embedded agents that use temporal-difference learning to make action recommendations. Finally, we propose a new area of research called "agency foundations" and pose four initial topics designed to improve our understanding of agency in AI-human interactions: benevolent game theory, algorithmic foundations of human rights, mechanistic interpretability of agency representation in neural-networks and reinforcement learning from internal states.
- Oceania > Australia (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- (5 more...)
- Information Technology (1.00)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Government (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- (2 more...)
The Case Against Explainability
Rozen, Hofit Wasserman, Elkin-Koren, Niva, Gilad-Bachrach, Ran
As artificial intelligence (AI) becomes more prevalent there is a growing demand from regulators to accompany decisions made by such systems with explanations. However, a persistent gap exists between the need to execute a meaningful right to explanation vs. the ability of Machine Learning systems to deliver on such a legal requirement. The regulatory appeal towards "a right to explanation" of AI systems can be attributed to the significant role of explanations, part of the notion called reason-giving, in law. Therefore, in this work we examine reason-giving's purposes in law to analyze whether reasons provided by end-user Explainability can adequately fulfill them. We find that reason-giving's legal purposes include: (a) making a better and more just decision, (b) facilitating due-process, (c) authenticating human agency, and (d) enhancing the decision makers' authority. Using this methodology, we demonstrate end-user Explainabilty's inadequacy to fulfil reason-giving's role in law, given reason-giving's functions rely on its impact over a human decision maker. Thus, end-user Explainability fails, or is unsuitable, to fulfil the first, second and third legal function. In contrast we find that end-user Explainability excels in the fourth function, a quality which raises serious risks considering recent end-user Explainability research trends, Large Language Models' capabilities, and the ability to manipulate end-users by both humans and machines. Hence, we suggest that in some cases the right to explanation of AI systems could bring more harm than good to end users. Accordingly, this study carries some important policy ramifications, as it calls upon regulators and Machine Learning practitioners to reconsider the widespread pursuit of end-user Explainability and a right to explanation of AI systems.
- Asia > Middle East > Israel > Tel Aviv District > Tel Aviv (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Asia > Middle East > Jordan (0.04)
- Overview (0.68)
- Research Report (0.50)
- Information Technology > Security & Privacy (1.00)
- Law > Statutes (0.68)
How Dictators Will Use Artificial Intelligence
Russia's savage, imperialistic and childish war on Ukraine has been said by democracies to be a battle between democracies and autocracies, the free world and the the unfree. And it is the opening of the battle to come between two very different sets of values. The other, more subtle, nefarious, insidious and perhaps deadlier in some ways, war is that of Artificial Intelligence. The abuse of AI has the capability to destroy human agency, take away any sense of free will, devastate human rights, divide societies and turn people under its thumb into automatons to serve the elites of corrupt, autocratic and dictatorial countries. To see how autocracies will use AI to subjugate and destroy any sense of human agency in their populations, we only have to look at how they've done so with social media.
The Future of Human Agency
This report covers results from the 15th "Future of the Internet" canvassing that Pew Research Center and Elon University's Imagining the Internet Center have conducted together to gather expert views about important digital issues. This is a nonscientific canvassing based on a nonrandom sample; this broad array of opinions about the potential influence of current trends may lead between 2022 and 2035 represents only the points of view of the individuals who responded to the queries. Pew Research Center and Elon's Imagining the Internet Center sampled from a database of experts to canvass from a wide range of fields, inviting entrepreneurs, professionals and policy people based in government bodies, nonprofits and foundations, technology businesses and think tanks, as well as interested academics and technology innovators. The predictions reported here came in response to a set of questions in an online canvassing conducted between June 29 and Aug. 8, 2022. In all, 540 technology innovators and developers, business and policy leaders, researchers and activists responded in some way to the question covered in this report. More on the methodology underlying this canvassing and the participants can be found in the section titled "About this canvassing of experts." Advances in the internet, artificial intelligence (AI) and online applications have allowed humans to vastly expand their capabilities and increase their capacity to tackle complex problems. These advances have given people the ability to instantly access and share knowledge and amplified their personal and collective power to understand and shape their surroundings. Today there is general agreement that smart machines, bots and systems powered mostly by machine learning and artificial intelligence will quickly increase in speed and sophistication between now and 2035.
- North America > United States > North Carolina > Orange County > Chapel Hill (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- (5 more...)
- Law (1.00)
- Information Technology > Security & Privacy (0.46)
Who Wrote this? How Smart Replies Impact Language and Agency in the Workplace
AI-mediated communication is designed to help us do our work more quickly and efficiently. But does it come at a cost? This study uses smart replies (SRs) to show how AI influences humans without any intent on the part of the developer - the very use of AI is sufficient. I propose a loss of agency theory as a viable approach for studying the impact of AI on human agency. This theory focusses on the transfer of agency that is forced by circumstances (such as time pressure), human weaknesses (such as complacency), and conceptual priming. Mixed methods involving a crowdsourced experiment test that theory. The quantitative results reveal that machine agency affects the content we author and the behavior we generate. But it is a non-zero-sum game. The transfers between human and machine agency are fluid; they complement, replace, and reinforce each other at the same time.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > Alaska > Anchorage Municipality > Anchorage (0.04)
- Asia > Japan > Honshū > Kantō > Kanagawa Prefecture > Yokohama (0.04)
- (5 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.93)
- Health & Medicine (0.68)
- Information Technology (0.68)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (0.88)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (0.48)
- Information Technology > Communications > Social Media > Crowdsourcing (0.34)