Black Box Deployed -- Functional Criteria for Artificial Moral Agents in the LLM Era
–arXiv.org Artificial Intelligence
The advancement of powerful yet opaque large language models (LLMs) necessitates a fundamental revision of the philosophical criteria used to evaluate artificial moral agents (AMAs). Pre-LLM frameworks often relied on the assumption of transparent architectures, which LLMs defy due to their stochastic outputs and opaque internal states. This paper argues that traditional ethical criteria are pragmatically obsolete for LLMs due to this mismatch. Engaging with core themes in the philosophy of technology, this paper proffers a revised set of ten functional criteria to evaluate LLM-based artificial moral agents: moral concordance, context sensitivity, normative integrity, metaethical awareness, system resilience, trustworthiness, corrigibility, partial transparency, functional autonomy, and moral imagination. These guideposts, applied to what we term "SMA-LLS" (Simulating Moral Agency through Large Language Systems), aim to steer AMAs toward greater alignment and beneficial societal integration in the coming years. We illustrate these criteria using hypothetical scenarios involving an autonomous public bus (APB) to demonstrate their practical applicability in morally salient contexts.
arXiv.org Artificial Intelligence
Jul-29-2025
- Country:
- Europe
- Estonia (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- North America > United States
- California
- San Diego County > San Diego (0.04)
- Santa Clara County > Palo Alto (0.04)
- Illinois > Cook County
- Chicago (0.04)
- California
- Europe
- Genre:
- Research Report (1.00)
- Industry:
- Government (1.00)
- Health & Medicine (1.00)
- Law (0.67)
- Technology: