security vulnerability
A social network for AI looks disturbing, but it's not what you think
A social network for AI looks disturbing, but it's not what you think A social network solely for AI - no humans allowed - has made headlines around the world. Chatbots are using it to discuss humans' diary entries, describe existential crises or even plot world domination . It looks like an alarming development in the rise of the machines - but all is not as it seems. Like any chatbots, the AI agents on Moltbook are just creating statistically plausible strings of words - there is no understanding, intent or intelligence. And in any case, there's plenty of evidence that much of what we can read on the site is actually written by humans.
- North America > United States > Maryland > Baltimore (0.05)
- Europe > United Kingdom > England > West Midlands > Birmingham (0.05)
- Europe > United Kingdom > England > Surrey (0.05)
- Health & Medicine (0.98)
- Information Technology > Services (0.95)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.74)
Chatbots Are Becoming Really, Really Good Criminals
Cybersecurity was already a nightmare. Listen to more stories on the Noa app. Earlier this fall, a team of security experts at the AI company Anthropic uncovered an elaborate cyber-espionage scheme. Hackers--strongly suspected by Anthropic to be working on behalf of the Chinese government--targeted government agencies and large corporations around the world. And it appears that they used Anthropic's own AI product, Claude Code, to do most of the work.
- Asia > China (0.47)
- North America > United States (0.05)
- Information Technology > Security & Privacy (1.00)
- Government > Military > Cyberwarfare (0.92)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.68)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.31)
Explaining Software Vulnerabilities with Large Language Models
Johnson, Oshando, Fomina, Alexandra, Krishnamurthy, Ranjith, Chaudhari, Vaibhav, Shanmuganathan, Rohith Kumar, Bodden, Eric
Abstract--The prevalence of security vulnerabilities has prompted companies to adopt static application security testing (SAST) tools for vulnerability detection. Nevertheless, these tools frequently exhibit usability limitations, as their generic warning messages do not sufficiently communicate important information to developers, resulting in misunderstandings or oversight of critical findings. In light of recent developments in Large Language Models (LLMs) and their text generation capabilities, our work investigates a hybrid approach that uses LLMs to tackle the SAST explainability challenges. In this paper, we present SAFE, an Integrated Development Environment (IDE) plugin that leverages GPT -4o to explain the causes, impacts, and mitigation strategies of vulnerabilities detected by SAST tools. Our expert user study findings indicate that the explanations generated by SAFE can significantly assist beginner to intermediate developers in understanding and addressing security vulnerabilities, thereby improving the overall usability of SAST tools. With the rise in software security vulnerabilities such as those in the Common Weakness Enumeration (CWE) Top 25 Most Dangerous Software Weaknesses list [1], many companies resort to static application security testing (SAST) tools for the detection of software vulnerabilities.
- North America > United States > California (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Germany > North Rhine-Westphalia (0.04)
- (2 more...)
Enhancing Cloud Security through Topic Modelling
Saleh, Sabbir M., Madhavji, Nazim, Steinbacher, John
Protecting cloud applications is critical in an era where security threats are increasingly sophisticated and persistent. Continuous Integration and Continuous Deployment (CI/CD) pipelines are particularly vulnerable, making innovative security approaches essential. This research explores the application of Natural Language Processing (NLP) techniques, specifically Topic Modelling, to analyse security-related text data and anticipate potential threats. We focus on Latent Dirichlet Allocation (LDA) and Probabilistic Latent Semantic Analysis (PLSA) to extract meaningful patterns from data sources, including logs, reports, and deployment traces. Using the Gensim framework in Python, these methods categorise log entries into security-relevant topics (e.g., phishing, encryption failures). The identified topics are leveraged to highlight patterns indicative of security issues across CI/CD's continuous stages (build, test, deploy). This approach introduces a semantic layer that supports early vulnerability recognition and contextual understanding of runtime behaviours.
- North America > Canada > Ontario > Middlesex County > London (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- (3 more...)
Security Degradation in Iterative AI Code Generation -- A Systematic Analysis of the Paradox
Shukla, Shivani, Joshi, Himanshu, Syed, Romilla
The rapid adoption of Large Language Models(LLMs) for code generation has transformed software development, yet little attention has been given to how security vulnerabilities evolve through iterative LLM feedback. This paper analyzes security degradation in AI-generated code through a controlled experiment with 400 code samples across 40 rounds of "improvements" using four distinct prompting strategies. Our findings show a 37.6% increase in critical vulnerabilities after just five iterations, with distinct vulnerability patterns emerging across different prompting approaches. This evidence challenges the assumption that iterative LLM refinement improves code security and highlights the essential role of human expertise in the loop. We propose practical guidelines for developers to mitigate these risks, emphasizing the need for robust human validation between LLM iterations to prevent the paradoxical introduction of new security issues during supposedly beneficial code "improvements".
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.89)
LLaVul: A Multimodal LLM for Interpretable Vulnerability Reasoning about Source Code
Jararweh, Ala, Adams, Michael, Sahu, Avinash, Mueen, Abdullah, Anwar, Afsah
Increasing complexity in software systems places a growing demand on reasoning tools that unlock vulnerabilities manifest in source code. Many current approaches focus on vulnerability analysis as a classifying task, oversimplifying the nuanced and context-dependent real-world scenarios. Even though current code large language models (LLMs) excel in code understanding, they often pay little attention to security-specific reasoning. We propose LLaVul, a multimodal LLM tailored to provide fine-grained reasoning about code through question-answering (QA). Our model is trained to integrate paired code and natural queries into a unified space, enhancing reasoning and context-dependent insights about code vulnerability. To evaluate our model performance, we construct a curated dataset of real-world vulnerabilities paired with security-focused questions and answers. Our model outperforms state-of-the-art general-purpose and code LLMs in the QA and detection tasks. We further explain decision-making by conducting qualitative analysis to highlight capabilities and limitations. By integrating code and QA, LLaVul enables more interpretable and security-focused code understanding.
- North America > United States > New Mexico (0.04)
- Asia > Singapore (0.04)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- (2 more...)
GitHub's Copilot Code Review: Can AI Spot Security Flaws Before You Commit?
As software development practices increasingly adopt AI-powered tools, ensuring that such tools can support secure coding has become critical. This study evaluates the effectiveness of GitHub Copilot's recently introduced code review feature in detecting security vulnerabilities. Using a curated set of labeled vulnerable code samples drawn from diverse open-source projects spanning multiple programming languages and application domains, we systematically assessed Copilot's ability to identify and provide feedback on common security flaws. Contrary to expectations, our results reveal that Copilot's code review frequently fails to detect critical vulnerabilities such as SQL injection, cross-site scripting (XSS), and insecure deserialization. Instead, its feedback primarily addresses low-severity issues, such as coding style and typographical errors. These findings expose a significant gap between the perceived capabilities of AI-assisted code review and its actual effectiveness in supporting secure development practices. Our results highlight the continued necessity of dedicated security tools and manual code audits to ensure robust software security.
- North America > United States (0.04)
- North America > Canada > Ontario > Toronto (0.04)
Firefox 143 gets Microsoft Copilot AI and Google Lens support
When you purchase through links in our articles, we may earn a small commission. The newest version of Firefox comes with new AI features and accessibility improvements, plus fixes to numerous security vulnerabilities. The latest update to Firefox brings the browser up to version 143 with various new features and improvements, including some that other browsers already offer. However, some of these features--like Google Lens--are only being introduced gradually. Mozilla plans to release Firefox 144 on October 14th, 2025.
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.86)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.67)
Google's in-house AI agent discovers critical vulnerability in Chrome
Google has fixed a critical vulnerability in Chrome versions 139.0.7258.154/155 According to Google, the vulnerability has not yet been exploited for attacks in the wild. The manufacturers of other Chromium-based browsers are expected to follow suit in the coming days. In the Chrome Releases blog post, Krishna Govind presents the eliminated vulnerability (CVE-2025-9478), which is treated as if it were discovered by external security researchers, but Google Big Sleep is named as the discoverer of the vulnerability. This is an "AI" tool based on Gemini for detecting security vulnerabilities and it's designed to detect vulnerabilities on its own without human assistance.
Assessing the Quality and Security of AI-Generated Code: A Quantitative Analysis
Sabra, Abbas, Schmitt, Olivier, Tyler, Joseph
This study presents a quantitative evaluation of the code quality and security of five prominent Large Language Models (LLMs): Claude Sonnet 4, Claude 3.7 Sonnet, GPT-4o, Llama 3.2 90B, and OpenCoder 8B. While prior research has assessed the functional performance of LLM-generated code, this research tested LLM output from 4,442 Java coding assignments through comprehensive static analysis using SonarQube. The findings suggest that although LLMs can generate functional code, they also introduce a range of software defects, including bugs, security vulnerabilities, and code smells. These defects do not appear to be isolated; rather, they may represent shared weaknesses stemming from systemic limitations within current LLM code generation methods. In particular, critically severe issues, such as hard-coded passwords and path traversal vulnerabilities, were observed across multiple models. These results indicate that LLM-generated code requires verification in order to be considered production-ready. This study found no direct correlation between a model's functional performance (measured by Pass@1 rate of unit tests) and the overall quality and security of its generated code, measured by the number of SonarQube issues in benchmark solutions that passed the functional tests. This suggests that functional benchmark performance score is not a good indicator of overall code quality and security. The goal of this study is not to rank LLM performance but to highlight that all evaluated models appear to share certain weaknesses. Consequently, these findings support the view that static analysis can be a valuable instrument for detecting latent defects and an important safeguard for organizations that deploy AI in software development.