Click any tag below to further narrow down your results
Links
This article discusses vulnerabilities in large language model (LLM) frameworks, highlighting specific case studies of security issues like remote code execution and SQL injection. It offers lessons learned for both users and developers, emphasizing the importance of validation and cautious implementation practices.
Augustus is a new security testing tool designed to identify vulnerabilities in large language models (LLMs), focusing on prompt injection and other attack vectors. Built in Go, it offers faster execution and lower memory usage compared to its Python-based predecessors. With over 210 vulnerability probes, it helps operators assess the security of various LLM providers efficiently.
qqqa is a command-line interface tool that combines two functions: asking questions and executing commands. It operates statelessly, allowing for quick interactions with various LLM providers like OpenAI and Claude without saving session history. The tool emphasizes security and ease of use, making it suitable for integration into existing shell workflows.
This article describes a framework for testing how AI models, specifically Opus 4.5 and GPT-5.2, generate exploits from vulnerability reports. It focuses on the experiments conducted using a QuickJS vulnerability, outlining the agents' strategies to bypass various security mitigations and achieve their objectives.
Datadog developed an LLM-powered tool called BewAIre to review pull requests for malicious activity in real time. The system processes code changes and classifies them, achieving over 99.3% accuracy while minimizing false positives. It addresses the challenges posed by the increasing volume of PRs and the sophistication of attacks.
Mercari's AI Security team created the LLM Key Server to streamline access to LLM APIs. This service allows users to obtain temporary API keys without manual requests, enhancing security while simplifying access for developers and non-developers alike.
The article explores using web browsers as a secure environment for running untrusted code, focusing on the potential of browser-based tools like Co-do. It discusses the importance of file and network isolation in maintaining user control and safety when executing code from sources like LLMs. The author highlights existing browser capabilities and suggests methods for improving sandboxing techniques.
This article outlines key security vulnerabilities identified by NVIDIA's AI Red Team in large language model (LLM) applications. It highlights risks such as remote code execution from LLM-generated code, insecure access in retrieval-augmented generation, and data exfiltration through active content rendering. The blog offers practical mitigation strategies for these issues.
The article discusses the potential security risks associated with using large language models (LLMs) in coding practices. It highlights how these models can inadvertently introduce vulnerabilities and the implications for developers and organizations. The need for robust security measures when integrating LLMs into coding workflows is emphasized.
Supabase's Model Context Protocol (MCP) poses a security risk as it can be exploited to leak sensitive SQL database information through user-submitted messages that are processed as commands. The integration allows developers to unintentionally execute harmful SQL queries due to elevated access privileges, emphasizing the need for better safeguards against prompt injection attacks.
Security backlogs often become overwhelming due to inconsistent severity labeling from various tools, leading to chaos in issue prioritization. Large language models (LLMs) can help by analyzing and scoring issues based on detailed context rather than relying solely on scanner outputs, providing a more informed approach to triage and prioritization.
Bloomberg's research reveals that the implementation of Retrieval-Augmented Generation (RAG) systems can unexpectedly increase the likelihood of large language models (LLMs) providing unsafe responses to harmful queries. The study highlights the need for enterprises to rethink their safety architectures and develop domain-specific guardrails to mitigate these risks.
Agentic AI systems, particularly those utilizing large language models (LLMs), face significant security vulnerabilities due to their inability to distinguish between instructions and data. The concept of the "Lethal Trifecta" highlights the risks associated with sensitive data access, untrusted content, and external communication, emphasizing the need for strict mitigations to minimize these threats. Developers must adopt careful practices, such as using controlled environments and minimizing data exposure, to enhance security in the deployment of these AI applications.
The article explores the concept of developing C2-less malware using large language models (LLMs) for autonomous decision-making and exploitation. It discusses the implications of such technology, particularly through a malware example called "PromptLock," which utilizes LLMs to generate and execute code without human intervention. The author proposes a proof of concept for self-contained malware capable of exploiting misconfigured services on a target system.
PromptMe is an educational project that highlights security vulnerabilities in large language model (LLM) applications, featuring 10 hands-on challenges based on the OWASP LLM Top 10. Aimed at AI security professionals, it provides a platform to explore risks and mitigation strategies, using Python and the Ollama framework. Users can set up the application to learn about vulnerabilities through CTF-style challenges, with solutions available for beginners.