safety engineering
What Is AI Safety? What Do We Want It to Be?
Harding, Jacqueline, Kirk-Giannini, Cameron Domenico
The field of AI safety seeks to prevent or reduce the harms caused by AI systems. A simple and appealing account of what is distinctive of AI safety as a field holds that this feature is constitutive: a research project falls within the purview of AI safety just in case it aims to prevent or reduce the harms caused by AI systems. Call this appealingly simple account The Safety Conception of AI safety. Despite its simplicity and appeal, we argue that The Safety Conception is in tension with at least two trends in the ways AI safety researchers and organizations think and talk about AI safety: first, a tendency to characterize the goal of AI safety research in terms of catastrophic risks from future systems; second, the increasingly popular idea that AI safety can be thought of as a branch of safety engineering. Adopting the methodology of conceptual engineering, we argue that these trends are unfortunate: when we consider what concept of AI safety it would be best to have, there are compelling reasons to think that The Safety Conception is the answer. Descriptively, The Safety Conception allows us to see how work on topics that have historically been treated as central to the field of AI safety is continuous with work on topics that have historically been treated as more marginal, like bias, misinformation, and privacy. Normatively, taking The Safety Conception seriously means approaching all efforts to prevent or mitigate harms from AI systems based on their merits rather than drawing arbitrary distinctions between them.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.05)
- North America > United States (0.04)
- Africa > Eswatini > Manzini > Manzini (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (1.00)
- Media (0.66)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.68)
From Hazard Identification to Controller Design: Proactive and LLM-Supported Safety Engineering for ML-Powered Systems
Hong, Yining, Timperley, Christopher S., Kästner, Christian
Machine learning (ML) components are increasingly integrated into software products, yet their complexity and inherent uncertainty often lead to unintended and hazardous consequences, both for individuals and society at large. Despite these risks, practitioners seldom adopt proactive approaches to anticipate and mitigate hazards before they occur. Traditional safety engineering approaches, such as Failure Mode and Effects Analysis (FMEA) and System Theoretic Process Analysis (STPA), offer systematic frameworks for early risk identification but are rarely adopted. This position paper advocates for integrating hazard analysis into the development of any ML-powered software product and calls for greater support to make this process accessible to developers. By using large language models (LLMs) to partially automate a modified STPA process with human oversight at critical steps, we expect to address two key challenges: the heavy dependency on highly experienced safety engineering experts, and the time-consuming, labor-intensive nature of traditional hazard analysis, which often impedes its integration into real-world development workflows. We illustrate our approach with a running example, demonstrating that many seemingly unanticipated issues can, in fact, be anticipated.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > New York (0.04)
- North America > United States > Florida > Palm Beach County > Boca Raton (0.04)
- Europe > Portugal > Lisbon > Lisbon (0.04)
- Transportation > Air (0.46)
- Information Technology > Security & Privacy (0.46)
Redefining Safety for Autonomous Vehicles
Koopman, Philip, Widen, William
Existing definitions and associated conceptual frameworks for computer-based system safety should be revisited in light of real-world experiences from deploying autonomous vehicles. Current terminology used by industry safety standards emphasizes mitigation of risk from specifically identified hazards, and carries assumptions based on human-supervised vehicle operation. Operation without a human driver dramatically increases the scope of safety concerns, especially due to operation in an open world environment, a requirement to self-enforce operational limits, participation in an ad hoc sociotechnical system of systems, and a requirement to conform to both legal and ethical constraints. Existing standards and terminology only partially address these new challenges. We propose updated definitions for core system safety concepts that encompass these additional considerations as a starting point for evolving safe-ty approaches to address these additional safety challenges. These results might additionally inform framing safety terminology for other autonomous system applications.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.05)
- North America > United States > Florida > Miami-Dade County > Miami (0.04)
- (3 more...)
- Transportation > Ground > Road (1.00)
- Law (1.00)
- Government > Military (1.00)
- (4 more...)
ChatSOS: LLM-based knowledge Q&A system for safety engineering
Tang, Haiyang, Liu, Zhenyi, Chen, Dongping, Chu, Qingzhao
Recent advancements in large language models (LLMs) have notably propelled natural language processing (NLP) capabilities, demonstrating significant potential in safety engineering applications. Despite these advancements, LLMs face constraints in processing specialized tasks, attributed to factors such as corpus size, input processing limitations, and privacy concerns. Obtaining useful information from reliable sources in a limited time is crucial for LLM. Addressing this, our study introduces an LLM-based Q&A system for safety engineering, enhancing the comprehension and response accuracy of the model. We employed prompt engineering to incorporate external knowledge databases, thus enriching the LLM with up-to-date and reliable information. The system analyzes historical incident reports through statistical methods, utilizes vector embedding to construct a vector database, and offers an efficient similarity-based search functionality. Our findings indicate that the integration of external knowledge significantly augments the capabilities of LLM for in-depth problem analysis and autonomous task assignment. It effectively summarizes accident reports and provides pertinent recommendations. This integration approach not only expands LLM applications in safety engineering but also sets a precedent for future developments towards automation and intelligent systems.
Physicist Max Tegmark on the promise and pitfalls of artificial intelligence
To describe Max Tegmark's career as "storied" is to do the Swedish-American physicist a disservice. He's published more than 200 publications and developed data analysis tools for microwave background experiments. And he's been elected as a Fellow of the American Physical Society for his contributions to cosmology. In 2015, Elon Musk donated $10 million to FLI to advance research into the ethical, legal, and economic effects of AI systems. Tegmark's latest book, Life 3.0: Being Human in the Age of Artificial Intelligence, postulates that neural networks of the future may be able to redesign their own hardware and internal structure.
- North America > United States > California (0.14)
- Europe > Russia (0.14)
- Asia > Russia (0.14)
- (9 more...)
- Personal > Interview (0.48)
- Summary/Review (0.34)
- Government > Military (0.95)
- Law Enforcement & Public Safety (0.71)
A Psychopathological Approach to Safety Engineering in AI and AGI
Behzadan, Vahid, Munir, Arslan, Yampolskiy, Roman V.
The complexity of dynamics in AI techniques is already approaching that of complex adaptive systems, thus curtailing the feasibility of formal controllability and reachability analysis in the context of AI safety. It follows that the envisioned instances of Artificial General Intelligence (AGI) will also suffer from challenges of complexity. To tackle such issues, we propose the modeling of deleterious behaviors in AI and AGI as psychological disorders, thereby enabling the employment of psychopathological approaches to analysis and control of misbehaviors. Accordingly, we present a discussion on the feasibility of the psychopathological approaches to AI safety, and propose general directions for research on modeling, diagnosis, and treatment of psychological disorders in AGI.
- North America > United States > Kansas (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (1.00)
- Information Technology > Security & Privacy (0.94)
- Health & Medicine > Therapeutic Area > Neurology (0.69)
Being Human in the Age of Artificial Intelligence with Max Tegmark and Neil deGrasse Tyson
Artificial intelligence is growing at an astounding rate, but are we ready for the consequences? Cosmologist and MIT physics professor Max Tegmark guides us through the state of artificial intelligence today, and the many paths we might take in further developing this technology. This Frontiers Lecture, moderated by Neil deGrasse Tyson, took place in the Museum's Hayden Planetarium on January 8, 2018. Max Tegmark will be participating in the 2018 Isaac Asimov Memorial Debate happening next week at the Museum. The podcast of that event will be available on February 15. ANNOUNCER: It is my pleasure to welcome not one but two of our amazing AMNH curators who will be introducing our presenter for the evening. First up we have Frederick P. Rose director of the Hayden Planetarium, Neil deGrasse Tyson. NEIL DEGRASSE TYSON (Frederick P. Rose Director of the Hayden Planetarium): Welcome to the universe. I've just got to see that that show of hands again, is this the first time you've ever attended a Hayden program? We've been here for 60 years. We do this every month.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > California (0.04)
- Europe > United Kingdom (0.04)
- Leisure & Entertainment > Games (1.00)
- Health & Medicine (1.00)
- Government > Military (0.68)
- (2 more...)
- Information Technology > Artificial Intelligence > Robots (0.68)
- Information Technology > Communications > Social Media (0.64)
- Information Technology > Communications > Mobile (0.60)
- (2 more...)
Interview: Max Tegmark on Superintelligent AI, Cosmic Apocalypse, and Life 3.0
IEEE Spectrum: Last Friday you had a discussion about AI with Yann LeCun, one of the most important computer scientists working on AI. LeCun said that since we don't know what form a superintelligent AI would take, it's premature to start researching safety mechanisms to control it. Max Tegmark: Just because we don't know quite what will go wrong doesn't mean we shouldn't think about it. That's the basic idea of safety engineering: You think hard about what might go wrong to prevent it from happening. But when the leaders of the Apollo program carefully thought through everything that could go wrong when you sent a rocket with astronauts to the moon, they weren't being alarmist. They were doing precisely what ultimately led to the success of the mission.