AI Security · OWASP LLM Top 10
LLM Security
The OWASP LLM Top 10 Explained for Enterprise
LLM security, generative AI security, and GenAI security are three names for the same problem space and the same control set: protecting language-model applications against the attack patterns that don\u2019t exist in traditional software. This guide walks through the full OWASP LLM Top 10 with mitigations, covers data poisoning and model supply-chain risk specifically, and gives the practical defense stack that turns the Top 10 from a checklist into an operational program.
30-SECOND EXECUTIVE TAKEAWAY
- OWASP LLM Top 10 is the floor. Coverage of all ten is the baseline; exceeding it is what defines a real LLM security program.
- Three risks dominate real incidents. Prompt injection, excessive agency, and improper output handling drive almost every disclosed LLM compromise.
- Architecture is the strongest control. Constrained tool permissions, default-deny on actions, and treating model output as untrusted reduce blast radius more than any input filter ever will.
Why LLM security needs its own checklist
OWASP\u2019s original Top 10 was built for web applications and the bug classes that show up in code. It still applies to the parts of an LLM system that look like a web app: the API in front of the model, the auth on the API, the database the application talks to. What it doesn\u2019t cover is the model itself, which has bug classes the original Top 10 was never designed for.
That\u2019s the gap the OWASP LLM Top 10 fills. The list is consensus-built across vendors, researchers, and enterprise security teams; the 2025 update reflected another year of real-world incidents and the rise of agentic AI. Most enterprises in 2026 use the LLM Top 10 as the engineering-side checklist alongside whatever they\u2019re doing for traditional application security.
The full list is below with practical mitigations for each. Coverage of all ten is the floor. The real work is making the controls actually run on the systems you have in production, with a named owner, a quarterly review, and a feedback loop into the AI risk register.
OWASP LLM TOP 10 (2025)
The full list, with mitigations
Each entry includes the threat in plain language and the practical mitigation enterprises apply in production. The numbering reflects OWASP\u2019s 2025 ordering by severity and prevalence.
Prompt Injection
User-supplied or third-party content that overrides developer intent.
Mitigation: Architecture controls; input filtering; output validation. See the dedicated guide.
Sensitive Information Disclosure
Models leak training data, system prompts, or context-window contents through crafted queries.
Mitigation: Limit context to what is needed; strip secrets from system prompts; output validation; PII detection on the outbound path.
Supply Chain
Compromised foundation models, fine-tuned weights, or model files (pickle, safetensors). Vulnerabilities inherited from the model layer.
Mitigation: Source models from trusted vendors only; scan model files for malicious payloads (ProtectAI, HiddenLayer); maintain an inventory and risk-tier of every model.
Data and Model Poisoning
Adversarial content in training, fine-tuning, or retrieval data that backdoors or biases the model.
Mitigation: Curate training and fine-tuning data; monitor RAG inputs; require provenance for any external data sources used in retrieval.
Improper Output Handling
Downstream systems trust model output without validation. Code execution, SQL injection, XSS via LLM output.
Mitigation: Treat LLM output as untrusted user input. Validate against schemas. Never pass to eval, exec, or shell. Render as plain text by default.
Excessive Agency
Agents with broad tool permissions take destructive or unauthorized actions when prompt-injected or hallucinating.
Mitigation: Default-deny tools. Allowlist per task. Human-in-the-loop on irreversible actions. Constrained permissions on each tool.
System Prompt Leakage
System prompts containing logic, examples, or secrets exposed to attackers.
Mitigation: Assume system prompts will leak. Don’t put secrets in them. Use them for instruction, not for security boundary.
Vector and Embedding Weaknesses
Embedding spaces leak information; access-control gaps in vector stores; embedding inversion attacks.
Mitigation: Apply per-tenant access control to vector stores. Avoid embedding sensitive data without need. Monitor for unusual query patterns.
Misinformation
Hallucinated outputs presented as fact; downstream decisions made on incorrect information.
Mitigation: Cite sources where possible. Validate with structured outputs against a known dataset. Set user expectations.
Unbounded Consumption
Resource exhaustion attacks: token-burning, denial-of-service, model wallet drain.
Mitigation: Rate limits per user and per session. Token budget enforcement. Anomaly detection on usage patterns.
Where data poisoning actually shows up
Data poisoning sounds dramatic and rarely is, in the form most enterprises expect. The 2024 OWASP update folded model poisoning into the same category to reflect that the operational risk for most enterprises lives in three pragmatic places: fine-tuning data quality (when the organization fine-tunes on data it didn\u2019t fully curate), retrieval data quality (RAG systems that ingest untrusted documents from email, third-party portals, or external collaborators), and feedback-loop data (RLHF or thumbs-up/down signals that an attacker can manipulate at scale).
The general defense pattern is the same across all three: provenance for every data source, monitoring for unusual content patterns in ingested data, and the ability to roll back a contaminated training run or a poisoned retrieval index. In 2026 most enterprises don\u2019t have those controls in place, which is why this category sits on the OWASP list and is likely to climb the ranking as more incidents surface.
FOR YOUR ROLE
What to do this quarter
For the technical CTO
Adopt the OWASP LLM Top 10 as the engineering checklist for every LLM-facing application. Run an inventory of which apps cover which risks. Require an architecture review before any new LLM feature ships, with the Top 10 on the checklist. Default-deny agent permissions across the board.
For the business CAIO
Translate the Top 10 into a one-page brief for the executive committee, with specific examples from your organization’s LLM deployments. Tie LLM security findings into the broader AI risk register. Fund the LLM gateway and red teaming layers of the security stack before they’re a board-level question. See the AI risk management guide.
For the CISO
Add OWASP LLM Top 10 patterns to red team and SOC playbooks. Build LLM-specific detection rules in the SIEM. Establish an incident response runbook for LLM compromise that covers context revocation, agent session kill, and forensic preservation.
LLM Security: Frequently Asked Questions
What is LLM security?
What is the OWASP LLM Top 10?
What is data poisoning in the context of LLMs?
What is excessive agency, and why is it on the OWASP list?
How is generative AI security different from LLM security?
Which OWASP LLM Top 10 risk is most exploited in production?
What’s the practical mitigation stack for LLM security?
Continue the AI security cluster
The Top 10 defines the surface; the rest of the cluster covers how to defend it.