privilege
LLM Agents Should Employ Security Principles
Zhang, Kaiyuan, Su, Zian, Chen, Pin-Yu, Bertino, Elisa, Zhang, Xiangyu, Li, Ninghui
Large Language Model (LLM) agents show considerable promise for automating complex tasks using contextual reasoning; however, interactions involving multiple agents and the system's susceptibility to prompt injection and other forms of context manipulation introduce new vulnerabilities related to privacy leakage and system exploitation. This position paper argues that the well-established design principles in information security, which are commonly referred to as security principles, should be employed when deploying LLM agents at scale. Design principles such as defense-in-depth, least privilege, complete mediation, and psychological acceptability have helped guide the design of mechanisms for securing information systems over the last five decades, and we argue that their explicit and conscientious adoption will help secure agentic systems. To illustrate this approach, we introduce AgentSandbox, a conceptual framework embedding these security principles to provide safeguards throughout an agent's life-cycle. We evaluate with state-of-the-art LLMs along three dimensions: benign utility, attack utility, and attack success rate. AgentSandbox maintains high utility for its intended functions under both benign and adversarial evaluations while substantially mitigating privacy risks. By embedding secure design principles as foundational elements within emerging LLM agent protocols, we aim to promote trustworthy agent ecosystems aligned with user privacy expectations and evolving regulatory requirements.
- North America > United States > Florida > Miami-Dade County > Miami (0.04)
- Asia > Middle East > Jordan (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (1.00)
Prompt Flow Integrity to Prevent Privilege Escalation in LLM Agents
Kim, Juhee, Choi, Woohyuk, Lee, Byoungyoung
Large Language Models (LLMs) are combined with plugins to create powerful LLM agents that provide a wide range of services. Unlike traditional software, LLM agent's behavior is determined at runtime by natural language prompts from either user or plugin's data. This flexibility enables a new computing paradigm with unlimited capabilities and programmability, but also introduces new security risks, vulnerable to privilege escalation attacks. Moreover, user prompt is prone to be interpreted in an insecure way by LLM agents, creating non-deterministic behaviors that can be exploited by attackers. To address these security risks, we propose Prompt Flow Integrity (PFI), a system security-oriented solution to prevent privilege escalation in LLM agents. Analyzing the architectural characteristics of LLM agents, PFI features three mitigation techniques -- i.e., untrusted data identification, enforcing least privilege on LLM agents, and validating unsafe data flows. Our evaluation result shows that PFI effectively mitigates privilege escalation attacks while successfully preserving the utility of LLM agents.
- Information Technology > Security & Privacy (1.00)
- Energy > Oil & Gas (0.67)
SymGPT: Auditing Smart Contracts via Combining Symbolic Execution with Large Language Models
Xia, Shihao, He, Mengting, Shao, Shuai, Yu, Tingting, Zhang, Yiying, Song, Linhai
To govern smart contracts running on Ethereum, multiple Ethereum Request for Comment (ERC) standards have been developed, each having a set of rules to guide the behaviors of smart contracts. Violating the ERC rules could cause serious security issues and financial loss, signifying the importance of verifying smart contracts follow ERCs. Today's practices of such verification are to manually audit each single contract, use expert-developed program-analysis tools, or use large language models (LLMs), all of which are far from effective in identifying ERC rule violations. This paper introduces SymGPT, a tool that combines the natural language understanding of large language models (LLMs) with the formal guarantees of symbolic execution to automatically verify smart contracts' compliance with ERC rules. To develop SymGPT, we conduct an empirical study of 132 ERC rules from three widely used ERC standards, examining their content, security implications, and natural language descriptions. Based on this study, we design SymGPT by first instructing an LLM to translate ERC rules into a defined EBNF grammar. We then synthesize constraints from the formalized rules to represent scenarios where violations may occur and use symbolic execution to detect them. Our evaluation shows that SymGPT identifies 5,783 ERC rule violations in 4,000 real-world contracts, including 1,375 violations with clear attack paths for stealing financial assets, demonstrating its effectiveness. Furthermore, SymGPT outperforms six automated techniques and a security-expert auditing service, underscoring its superiority over current smart contract analysis methods.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Europe > Austria > Vienna (0.14)
- North America > United States > California > San Diego County > San Diego (0.04)
- (14 more...)
- Information Technology > Security & Privacy (1.00)
- Banking & Finance > Trading (1.00)
- Banking & Finance > Economy (1.00)
Privilege Scores
Bothmann, Ludwig, Boustani, Philip A., Alvarez, Jose M., Casalicchio, Giuseppe, Bischl, Bernd, Dandl, Susanne
Bias-transforming methods of fairness-aware machine learning aim to correct a non-neutral status quo with respect to a protected attribute (PA). Current methods, however, lack an explicit formulation of what drives non-neutrality. We introduce privilege scores (PS) to measure PA-related privilege by comparing the model predictions in the real world with those in a fair world in which the influence of the PA is removed. At the individual level, PS can identify individuals who qualify for affirmative action; at the global level, PS can inform bias-transforming policies. After presenting estimation methods for PS, we propose privilege score contributions (PSCs), an interpretation method that attributes the origin of privilege to mediating features and direct effects. We provide confidence intervals for both PS and PSCs. Experiments on simulated and real-world data demonstrate the broad applicability of our methods and provide novel insights into gender and racial privilege in mortgage and college admissions applications.
- Europe > Switzerland > Zürich > Zürich (0.14)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- North America > United States > Wisconsin (0.04)
- (6 more...)
- Law > Civil Rights & Constitutional Law (0.92)
- Education > Educational Setting > Higher Education (0.34)
AutoPenBench: Benchmarking Generative Agents for Penetration Testing
Gioacchini, Luca, Mellia, Marco, Drago, Idilio, Delsanto, Alexander, Siracusano, Giuseppe, Bifulco, Roberto
Generative AI agents, software systems powered by Large Language Models (LLMs), are emerging as a promising approach to automate cybersecurity tasks. Among the others, penetration testing is a challenging field due to the task complexity and the diverse strategies to simulate cyber-attacks. Despite growing interest and initial studies in automating penetration testing with generative agents, there remains a significant gap in the form of a comprehensive and standard framework for their evaluation and development. This paper introduces AutoPenBench, an open benchmark for evaluating generative agents in automated penetration testing. We present a comprehensive framework that includes 33 tasks, each representing a vulnerable system that the agent has to attack. Tasks are of increasing difficulty levels, including in-vitro and real-world scenarios. We assess the agent performance with generic and specific milestones that allow us to compare results in a standardised manner and understand the limits of the agent under test. We show the benefits of AutoPenBench by testing two agent architectures: a fully autonomous and a semi-autonomous supporting human interaction. We compare their performance and limitations. For example, the fully autonomous agent performs unsatisfactorily achieving a 21% Success Rate (SR) across the benchmark, solving 27% of the simple tasks and only one real-world task. In contrast, the assisted agent demonstrates substantial improvements, with 64% of SR. AutoPenBench allows us also to observe how different LLMs like GPT-4o or OpenAI o1 impact the ability of the agents to complete the tasks. We believe that our benchmark fills the gap with a standard and flexible framework to compare penetration testing agents on a common ground. We hope to extend AutoPenBench along with the research community by making it available under https://github.com/lucagioacchini/auto-pen-bench.
- Europe > Italy > Piedmont > Turin Province > Turin (0.04)
- Europe > Germany > Baden-Württemberg > Karlsruhe Region > Heidelberg (0.04)
- Information Technology > Security & Privacy (1.00)
- Government > Military > Cyberwarfare (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.54)
Towards Automated Penetration Testing: Introducing LLM Benchmark, Analysis, and Improvements
Isozaki, Isamu, Shrestha, Manil, Console, Rick, Kim, Edward
Hacking poses a significant threat to cybersecurity, inflicting billions of dollars in damages annually. To mitigate these risks, ethical hacking, or penetration testing, is employed to identify vulnerabilities in systems and networks. Recent advancements in large language models (LLMs) have shown potential across various domains, including cybersecurity. However, there is currently no comprehensive, open, end-to-end automated penetration testing benchmark to drive progress and evaluate the capabilities of these models in security contexts. This paper introduces a novel open benchmark for LLM-based automated penetration testing, addressing this critical gap. We first evaluate the performance of LLMs, including GPT-4o and Llama 3.1-405B, using the state-of-the-art PentestGPT tool. Our findings reveal that while Llama 3.1 demonstrates an edge over GPT-4o, both models currently fall short of performing fully automated, end-to-end penetration testing. Next, we advance the state-of-the-art and present ablation studies that provide insights into improving the PentestGPT tool. Our research illuminates the challenges LLMs face in each aspect of Pentesting, e.g. enumeration, exploitation, and privilege escalation. This work contributes to the growing body of knowledge on AI-assisted cybersecurity and lays the foundation for future research in automated penetration testing using large language models.
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- Europe (0.04)
- Information Technology > Security & Privacy (1.00)
- Government > Military > Cyberwarfare (0.76)
- Government > Regional Government > North America Government > United States Government (0.46)
AI Has Become a Technology of Faith
An important thing to realize about the grandest conversations surrounding AI is that, most of the time, everyone is making things up. This isn't to say that people have no idea what they're talking about or that leaders are lying. But the bulk of the conversation about AI's greatest capabilities is premised on a vision of a theoretical future. It is a sales pitch, one in which the problems of today are brushed aside or softened as issues of now, which surely, leaders in the field insist, will be solved as the technology gets better. What we see today is merely a shadow of what is coming.
- Health & Medicine (1.00)
- Information Technology > Security & Privacy (0.69)
Promoting Constructive Deliberation: Reframing for Receptiveness
Kambhatla, Gauri, Lease, Matthew, Rajadesingan, Ashwin
To promote constructive discussion of controversial topics online, we propose automatic reframing of disagreeing responses to signal receptiveness to a preceding comment. Drawing on research from psychology, communications, and linguistics, we identify six strategies for reframing. We automatically reframe replies to comments according to each strategy, using a Reddit dataset. Through human-centered experiments, we find that the replies generated with our framework are perceived to be significantly more receptive than the original replies and a generic receptiveness baseline. We illustrate how transforming receptiveness, a particular social science construct, into a computational framework, can make LLM generations more aligned with human perceptions. We analyze and discuss the implications of our results, and highlight how a tool based on our framework might be used for more teachable and creative content moderation.
- North America > United States > Alabama (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (8 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.94)
- Law (0.94)
- Government (0.93)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.46)
House of the Dragon season two: Dragons, grief and family feuds
House of the Dragon was one of the few US productions that continued to shoot during last year's Hollywood writers' and actors' strikes. For almost three months in 2023, industry writers and actors walked out in a dispute over fair pay and the use of artificial intelligence in the indsutry. But the House of the Dragon cast did not take part because the show was mainly filmed in the UK under contracts overseen by British union Equity, rather than its striking US counterpart the Screen Actors Guild (SAG). The show's writer Ryan Condal told the BBC it was a "fraught period", but a "great privilege" to keep the cast and crew employed. "There was lots of labour strife," he said. "We were lucky all of our scripts were done and we were already in production when when the writers' strike happened.
Sen. Tom Cotton torches Google AI system as 'racist, preposterously woke, Hamas-sympathizing'
Radio host Tommy Sotomayor reacts to artificial intelligence images rewriting history, on'Jesse Watters Primetime.' Sen. Tom Cotton, R-Ark., slammed Google's AI chatbot Gemini as "preposterously woke" on Friday for its refusal to produce any images of White people. The company paused the chatbot's image generation on Thursday after social media users pointed out that the system was creating inaccurate historical images that sometimes replaced White people, like the Founding Fathers, with images of Black, Native American and Asian people. "Google deserves condemnation for creating a racist, preposterously woke, Hamas-sympathizing AI system," Cotton said in a statement on X, formerly Twitter. "Republican lawmakers will remember this the next time Google comes asking for antitrust help."
- North America > United States (0.95)
- Asia > Middle East > Palestine (0.62)