Goto

Collaborating Authors

 catastrophic harm


What Is AI Safety? What Do We Want It to Be?

Harding, Jacqueline, Kirk-Giannini, Cameron Domenico

arXiv.org Artificial Intelligence

The field of AI safety seeks to prevent or reduce the harms caused by AI systems. A simple and appealing account of what is distinctive of AI safety as a field holds that this feature is constitutive: a research project falls within the purview of AI safety just in case it aims to prevent or reduce the harms caused by AI systems. Call this appealingly simple account The Safety Conception of AI safety. Despite its simplicity and appeal, we argue that The Safety Conception is in tension with at least two trends in the ways AI safety researchers and organizations think and talk about AI safety: first, a tendency to characterize the goal of AI safety research in terms of catastrophic risks from future systems; second, the increasingly popular idea that AI safety can be thought of as a branch of safety engineering. Adopting the methodology of conceptual engineering, we argue that these trends are unfortunate: when we consider what concept of AI safety it would be best to have, there are compelling reasons to think that The Safety Conception is the answer. Descriptively, The Safety Conception allows us to see how work on topics that have historically been treated as central to the field of AI safety is continuous with work on topics that have historically been treated as more marginal, like bias, misinformation, and privacy. Normatively, taking The Safety Conception seriously means approaching all efforts to prevent or mitigate harms from AI systems based on their merits rather than drawing arbitrary distinctions between them.


U.K. pushes nations to label AI as capable of 'catastrophic harm'

The Japan Times

British Prime Minister Rishi Sunak is pushing for nations to label artificial intelligence as capable of "catastrophic harm" at the AI Safety Summit the U.K. is hosting next month as it seeks to forge a common international approach on the rapidly advancing technology. Britain wants countries to sign up to a joint position that outlines particular concerns for AI's impact on cybersecurity and biotechnology, according to a draft communique circulated to attendees and seen by Bloomberg. Officials aim to hammer out final wording of the communique by Oct. 25, a separate document showed. "There is potential for significant, even catastrophic, harm, either deliberate or unintentional, stemming from the most dangerous capabilities of these AI models," according to the draft, dated Oct. 16.

  Country: Europe > United Kingdom (0.69)
  Industry: Government (0.69)