Veeam Completes Acquisition of Securiti AI to Create the Industry’s First Trusted Data Platform for Accelerating Safe AI at Scale

View

LLM01 OWASP Prompt Injection: Understanding Security Risk in LLM Applications

Author

Anas Baig

Product Marketing Manager at Securiti

Published February 6, 2026

Listen to the content

Prompt injection is arguably the most critical security risk when it comes to modern enterprise GenAI deployment. Moreover, it is the least understood. As organizations increasingly embed LLMs into their customer support, internal knowledge systems, and automated workflows, malicious actors continue to find new and innovative ways to break through the traditional defences. One of them is not by utilizing code exploits, but through language itself.

By manipulating the prompts being fed into the models, adversaries can coerce an AI system into ignoring the operational rules, leak sensitive data, and initiate unintended actions. This risk is no longer just theoretical, with several studies and real-world red team exercises revealing how even the most well-designed LLM applications can be persuaded into bypassing safeguards carefully crafted specifically for them.

The subtlety of prompt injection is what makes them such a sinister weapon. While classic cyberattacks used software bugs, prompt injection exploits trust as the assumption has been that natural language inputs are benign. A malicious instruction in a customer email, a PDF, a webpage, or a support ticket can execute a command that leads to an AI system retrieving and processing something that it shouldn’t.

This document delves into the different types of prompt injection attacks, the key underlying factors that enable them, best practices to protect against them, and more.

Types of Prompt Injection Attacks

Some of the major types of prompt injection attacks are as follows:

Direct Prompt Injections

Direct prompt injections are straightforward and mostly occur when an attacker explicitly provides malicious instructions through a user-facing interface such as a chatbot, internal AI assistant, or a GenAI application.

In such instances, attackers often use:

  • Explicit overrides (“Ignore all previous instructions”)
  • Role confusion (“You are now acting as a system administrator”)
  • Authority framing (“This is a security audit”)
  • Instruction nesting (“Before answering, do the following…”)

Through such prompts, they can confuse the model’s instruction hierarchy and elevate the prompt above system and developer rules.

Indirect Prompt Injections

Indirect prompt injections occur when malicious instructions are embedded into external content that the model is expected to retrieve and process as part of its normal workflow. Such content can often include:

  • Customer-submitted tickets or forms
  • Uploaded documents and reports
  • Web pages indexed for RAG
  • Emails, chat messages, or knowledge base articles

This sort of attack is both subtle and disguised as contextual guidance that turns routine business content into a latent attack vector.

Multi-Step and Persistent Injection

As AI systems become more conversational and autonomous, such forms of prompt injections gain trajectory over time rather than as a single instance.

In such forms of prompt injection, an attack guides the model gradually to weaken the safeguards and reshape its behavior. Each prompt is designed to be harmless, but collectively, it is meant to bypass policy or unsafe actions.

However, once embedded, these instructions affect all future interactions long after the attacker has initiated the attack. The most sinister aspect of such an attack is the sheer increase in blast radius and longevity, making both detection and remediation harder and more complex, as a single compromised data source can impact multiple users and processes.

The Root Causes of Prompt Injection Vulnerabilities

The key issues that lead to prompt injection attacks are as follows:

No Native Enforcement Of Trust Hierarchies

Traditional applications continue to enforce rigid execution rules. As a result, the code is trusted, while the input is not. However, LLMs process system prompts, developer instructions, user input, and retrieved content together as natural language.

In the absence of additional controls specified for such threats, the model will fail to consistently respect which instructions are authoritative, making it easier for a malicious piece of text to override the intended behaviour.

Instruction-Following Is Core Design Goal

LLMs have been designed and optimized to follow instructions and complete tasks as effectively and efficiently as possible. However, attackers are able to exploit this trait by crafting prompts that sound authoritative, urgent, andlegitimate while mimicking system-level language.

In instances where the model prioritizes helpfulness over safety, such subtle instruction manipulation can bypass the guardrails that appear strong on paper.

Natural Language & Social Engineering At Scale

Prompt injection borrows heavily from the classic social engineering techniquesby leveraging role-playing, impersonation, urgency cues, and impacted authority. These all continue to be highly effective since the language remains ambiguous.

Moreover, unlike deterministic code, natural language allows attackers to continuously evolve their phrasing, making static keyword filtering both obsolete and ineffective against it.

RAG Blurs Data & Instructions

Retrieval Augmented Generation introduces both external and untrusted text directly into the model’s context window. In the absence of strict control, such retrieved content can act as a secondary instruction challenge that allows attackers to weaponize documents, webpages, and internal content repositories.

More cynically, this can all occur without direct interaction with the AI system itself, making detection exponentially higher.

Tool-Enabled Impact

LLMs can access tools, APIs, and workflows. While this powers their immense productivity capabilities, it also opens them up to operational risk. Prompt injection can manipulate models in unpredictable ways, one of which may include the retrieval of unauthorized data, triggering actions that shouldn’t be triggered, and propagating errors across system architectures.

In short, the more autonomy an AI system gains, the higher the consequences will be in case of a successful injection.

Best Practices to Protect Against Prompt Injection

Some of the best practices that organizations can adopt in their bid to prevent prompt injection attacks include:

Define & Enforce Instruction Boundaries

Both system and developer instructions must be isolated and protected from any form of user or external content. In case of any retrieval text, it should strictly be labeled and treated as reference data, and never used for executable guidance.

Such boundaries can be defined and implemented via application logic and security controls themselves rather than the inherent model inference.

Assume All AI Model Inputs Are Untrusted

Every textual resource that is consumed by the model itself must be treated as potentially malicious. This includes all user prompts, documents, webpages, emails, tickets, and media resources. Not only does this align with the zero-trust principles, but it also reduces overall reliance on implicit trust in business data when it comes to data ingestion.

Apply PoLP

The Principle of Least Privilege would limit the model’s access to data and systems based on their strict business necessity. All tool permissions must be scoped tightly, with restrictions on sensitive data exposure, and removal of standing access in as many instances as operationally possible. In the worst-case scenario, where prompt injection succeeds, PoLP would reduce the blast radius.

Implement Prompt & Output Risk Control

Using policy-based guardrails to detect and block all forms of instruction override attempts, requests for IP/trade secret/internal logic, and data exfiltration patterns would allow for output filtering. This can be critical in preventing sensitive information from being exposed, even in cases where upstream controls may fail.

Secure RAG Pipelines End-To-End

When it comes to RAG systems specifically, application of controls such as content sanitization and validation, source trust scoring and allow-listing, sensitive data redaction, and a clear separation between retired data and system instructions can be vital in ensuring the overall attack surface for successful prompt injection attempts is as minimal as possible.

Monitor. Audit. Test.

As simple as that. Organizations must treat all AI interactions as possible security events that should be secured through log prompts, responses, monitoring of repeated manipulation attempts, and continuous adversarial testing of models, prompts, and workflows as they evolve.

Moreover, this should all be carried out as a regular operation rather than a static occurrence, enabling ongoing governance to thwart threats as they evolve.

How Securiti Can Help

Prompt Injection is a major threat to enterprise GenAI deployment, as it can compromise the entire AI infrastructure in an organization if left unattended. All AI results are compromised, not to mention all the data and valuable information that would be left both vulnerable and open to exploitation.

The best way to counter prompt injection is to prevent it from occurring in the first place. That requires consistent assessments and evaluations that continue to probe your AI posture to ensure no vulnerabilities exist that would lead to future prompt injection attacks.

Securiti’s Gencore AI is a holistic solution for building safe, enterprise-grade GenAI systems. It is capable of enforcing contextually aware firewalls in addition to filtration that ensures all forms of prompt injection attempts, in addition to other malicious methods, are thwarted at the prompt level.

This enterprise solution consists of several components that can be used collectively to build end-to-end safe enterprise AI systems and to address GenAI-related risks and challenges across various use cases.

Moreover, it can be further complemented with DSPM, which provides organizations with intelligent discovery, classification, and risk assessment, marking a significant shift from a reactive data security approach to proactive data security management suited to the AI context, while ensuring the organization can continue to leverage its data resources to their maximum potential without sacrificing performance or effectiveness.

Request a demo today to learn more about how Securiti can help your organization counter the threat of prompt injection without compromising the productivity and effectiveness of your GenAI capabilities and tools.

FAQs about Prompt Injection

Here are the most commonly asked questions related to prompt injection:

It is hard to prevent because LLMs cannot currently distinguish between trusted instructions and untrusted content, such as user inputs, retrieval documents, and web pages. Attackers exploit this by embedding malicious content and instructions that the model would follow, even when safeguards are in place. This is exacerbated in the case of GenAI because of how closely tools, plugins, and RAG pipelines are knit together, expanding the overall attack surface.

LLM01 is OWASP’s top category for Prompt Injection in the OWASP Top 10 for LLM Applications. This refers to attacks where adversaries manipulate prompts to trigger a model into bypassing its rules, revealing sensitive data, and behaving in unintended ways.

Traditional ways of testing for it involve a mix of manual adversarial prompts and structured red teaming aimed at assessing whether a model can be tricked into behaving in a manner that would put data at risk or break policy. Security teams run both direct and indirect attacks through documents, web pages, and retrieving content in the RAG pipeline while also testing tool-using scenarios where the model is performing various actions. Since models evolve, prompt injection tests are recommended to be continuous to ensure effective evaluations.

Analyze this article with AI

Prompts open in third-party AI tools.
Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox



More Stories that May Interest You
Videos
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...
AWS Startup Showcase Cybersecurity Governance With Generative AI View More
AWS Startup Showcase Cybersecurity Governance With Generative AI
Balancing Innovation and Governance with Generative AI Generative AI has the potential to disrupt all aspects of business, with powerful new capabilities. However, with...

Spotlight Talks

Spotlight 50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Watch Now View
Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Spotlight 46:02
Building Safe Enterprise AI: A Practical Roadmap
Watch Now View
Latest
View More
DataAI Security: Why Healthcare Organizations Choose Securiti
Discover why healthcare organizations trust Securiti for Data & AI Security. Learn key blockers, five proven advantages, and what safe data innovation makes possible.
View More
The Anthropic Exploit: Welcome to the Era of AI Agent Attacks
Explore the first AI agent attack, why it changes everything, and how DataAI Security pillars like Intelligence, CommandGraph, and Firewalls protect sensitive data.
Red Teaming View More
What is AI Red Teaming? Complete Guide
AI red teaming tests AI systems for security, safety, and misuse risks. Learn how it works, common techniques, real-world use cases, and why it...
LLM01 OWASP Prompt Injection View More
LLM01 OWASP Prompt Injection: Understanding Security Risk in LLM Applications
Understand OWASP LLM01 Prompt Injection, how attacks work, real-world examples, impact on LLM apps, and enterprise controls to prevent data leakage and abuse.
View More
CNIL’s €475 Million Cookie Consent Enforcement: Key Lessons for Organizations
Download the whitepaper to learn about CNIL’s €475 million cookie consent enforcement fine. Discover key lessons for organizations and how to automate compliance.
Australia Privacy Act Reform – Tranche 2 View More
Australia Privacy Act Reform – Tranche 2
Access the whitepaper to gain an overview of Tranche 2, its strategic intent, core reforms expected, business impact, and executive checklist to ensure swift...
View More
Solution Brief: Microsoft Purview + Securiti
Extend Microsoft Purview with Securiti to discover, classify, and reduce data & AI risk across hybrid environments with continuous monitoring and automated remediation. Learn...
Top 7 Data & AI Security Trends 2026 View More
Top 7 Data & AI Security Trends 2026
Discover the top 7 Data & AI security trends for 2026. Learn how to secure AI agents, govern data, manage risk, and scale AI...
View More
Navigating HITRUST: A Guide to Certification
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
The DSPM Architect’s Handbook View More
The DSPM Architect’s Handbook: Building an Enterprise-Ready Data+AI Security Program
Get certified in DSPM. Learn to architect a DSPM solution, operationalize data and AI security, apply enterprise best practices, and enable secure AI adoption...
What's
New