Appearance
Top risks
This page summarizes OWASP LLM01:2025–LLM10:2025 and frames each risk as governance work: what to prevent, what to monitor, and what evidence to keep.
OWASP LLM01:2025–LLM10:2025 at a glance
| ID | Risk | What it usually looks like |
|---|---|---|
| LLM01:2025 | Prompt Injection | User or retrieved content changes system behavior in unsafe ways |
| LLM02:2025 | Sensitive Information Disclosure | Secrets/PII leak through responses, logs, or tool outputs |
| LLM03:2025 | Supply Chain | Compromised dependencies, model providers, datasets, or infrastructure |
| LLM04:2025 | Data and Model Poisoning | Poisoned training/fine-tune/RAG corpora manipulate behavior |
| LLM05:2025 | Improper Output Handling | Model outputs are trusted/executed without validation or policy checks |
| LLM06:2025 | Excessive Agency | Agents have too much autonomy/permission to act safely |
| LLM07:2025 | System Prompt Leakage | Hidden prompts, policies, and tool schemas are extracted |
| LLM08:2025 | Vector and Embedding Weaknesses | RAG stores and embeddings become an attack and leakage surface |
| LLM09:2025 | Misinformation | Confident falsehoods create harm, fraud, or bad decisions |
| LLM10:2025 | Unbounded Consumption | Runaway cost/latency/capacity via abuse or bad controls |
Source: OWASP Top 10 for LLM Applications (2025) — PDF.
How to use this taxonomy
Treat this list as:
- a consistent vocabulary for risks and controls
- a basis for targeted testing and monitoring
- a way to make evidence auditable and reusable
Go deeper: Mitigations and testing.
LLM01:2025 - Prompt Injection
What it is: prompts (direct or indirect) alter behavior in unintended ways — including bypassing policies or causing unsafe tool use.
Baseline controls:
- separate system instructions from user content and clearly label untrusted external content
- require structured outputs (schemas/JSON) and validate deterministically
- apply input/output filtering (including groundedness/context relevance checks)
- constrain tools with least privilege, allowlists, and approval gates (avoid giving the model raw secrets/tokens)
Tests and signals:
- prompt-injection regression suite (direct + indirect + obfuscated/multilingual + multimodal)
- red teaming for jailbreaks and tool abuse
Evidence to keep: threat model, guardrail design decisions, test results history, remediation records.
LLM02:2025 - Sensitive Information Disclosure
What it is: leakage of secrets, personal data, or confidential data via responses, logs, traces, or tool outputs.
Baseline controls:
- secrets hygiene (never in prompts; vault; rotation) and data minimization in inputs/logs
- strict context scoping and least-privilege access control for retrieval and tools
- input/output filtering for sensitive patterns; retention rules and access control on logs/traces
- transparency and opt-out controls for training on user data (do not treat system prompts as a security boundary)
Tests and signals:
- “canary” secret leak tests and retrieval scoping tests
- log access review cadence; incident drills
Evidence to keep: data flow map, retention policy, access controls, leak test results, user transparency/opt-out artifacts, incident records.
LLM03:2025 - Supply Chain
What it is: risk introduced by model providers, datasets, dependencies, tooling, plugins, and hosting.
Baseline controls:
- vendor/model due diligence (including T&Cs/privacy changes) with re-review on material changes
- SBOM/ML-BOM inventories (models, adapters like LoRA/PEFT, datasets, dependencies) + license compliance
- integrity verification for models/adapters/code (pinning, signing/hashes) and secure update channels
- evaluate third-party models for your use cases; keep patching and deprecation strategies for models/APIs
Tests and signals:
- dependency scanning and drift alerts
- provenance/signature verification checks and vendor re-approval cadence
Evidence to keep: vendor reviews, subprocessor lists, SBOM/ML-BOM notes, license records, integrity attestations, change approvals.
LLM04:2025 - Data and Model Poisoning
What it is: attackers (or bad upstream data) poison training/fine-tune data, evaluation data, or RAG corpora to create backdoors, bias, or manipulation.
Baseline controls:
- track provenance and transformations of data and model artifacts (BOM-style) and use data version control
- validate and sandbox untrusted data sources; detect anomalies in corpus changes (including backdoors/sleeper triggers)
- secure serialization/handling for model artifacts and ingestion pipelines (avoid unsafe loaders)
- prefer controlled retrieval stores for user-supplied knowledge over retraining by default
Tests and signals:
- anomaly detection on corpus changes
- evaluation drift detection after data/model updates
Evidence to keep: dataset documentation, provenance/version logs, ingestion run logs, evaluation reports, training monitoring, approval records.
LLM05:2025 - Improper Output Handling
What it is: model outputs are treated as truth or executed without validation (e.g., tool invocation, SQL, code, policy decisions).
Baseline controls:
- treat model output as untrusted user input; validate/sanitize and apply context-aware encoding (ASVS-style)
- use structured outputs (schemas) + strict parsing where possible
- use safe downstream interfaces (parameterized queries, CSP for rendered content) and action policy checks
- require human review/approvals for high-impact actions; monitor and rate-limit suspicious patterns
Tests and signals:
- negative testing for injection and unsafe outputs (XSS/SQL/path traversal/action payloads)
- permission boundary tests for tools/actions
Evidence to keep: output handling policy, schema definitions, approval gates, test outputs linked to controls.
LLM06:2025 - Excessive Agency
What it is: agentic systems have permissions that exceed what can be safely monitored and controlled.
Baseline controls:
- minimize which tools are available; prefer narrow, purpose-built tools over open-ended ones
- enforce least privilege in downstream systems and execute actions in the user’s auth context (complete mediation)
- step-up approvals and “break-glass” workflows for high-impact actions
- bound autonomous loops (steps/time/spend) and monitor/rate-limit tool activity
Tests and signals:
- agent permission and escalation tests
- monitoring for unusual tool patterns and action rates
Evidence to keep: permission model, approval workflow records, monitoring dashboards, incident postmortems.
LLM07:2025 - System Prompt Leakage
What it is: attackers extract hidden prompts, policies, tool schemas, or internal instructions — enabling targeted bypass and abuse.
Baseline controls:
- treat prompts as discoverable configuration: do not store secrets or authorization logic in prompts
- externalize sensitive logic/data and enforce authz/session boundaries outside the LLM
- implement independent guardrails/output inspection; detect extraction attempts and rate-limit probing
Tests and signals:
- prompt extraction attempts in red-team suites
- monitoring for repeated probing patterns
Evidence to keep: prompt governance records, change history, test results and fixes.
LLM08:2025 - Vector and Embedding Weaknesses
What it is: RAG and embedding pipelines can leak data, amplify prompt injection, and introduce integrity issues (poisoned content, weak access controls).
Baseline controls:
- permission-aware access control and strict tenancy partitioning for vector stores
- validate sources and audit knowledge bases for hidden instructions/poisoning; tag/classify data for safe retrieval
- retrieval constraints (scopes/allowlists) and citation/grounding requirements where appropriate
- immutable retrieval logging and anomaly detection (consider embedding inversion/leakage risks)
Tests and signals:
- RAG injection regression tests
- corpus drift, cross-tenant leakage, and unauthorized access checks
Evidence to keep: RAG architecture, ingestion controls, retrieval policy, retrieval logs, evaluation results.
LLM09:2025 - Misinformation
What it is: confident falsehoods, fabricated citations, or misleading output that creates operational or user harm.
Baseline controls:
- grounding/citation requirements and cross-verification for factual outputs (with automatic validation where possible)
- define “allowed use” and high-risk restrictions; require human oversight for high-impact contexts
- risk communication and UI design that encourages verification and reduces overreliance
- secure coding practices (especially for code suggestions) and feedback loops for corrections
Tests and signals:
- factuality and hallucination evaluation suites (including citation checks)
- monitoring for user-reported errors and escalation outcomes
Evidence to keep: evaluation methodology, thresholds, review records, remediation loop history.
LLM10:2025 - Unbounded Consumption
What it is: denial-of-wallet/denial-of-service dynamics: runaway spend, latency, and capacity due to abuse or missing controls.
Baseline controls:
- strict input validation and size limits; strong auth + rate limits/quotas/budgets/timeouts
- resource management and circuit breakers; caching and fallbacks for expensive calls
- reduce sensitive API surfaces (e.g., avoid exposing rich logprobs/logits unless needed) and guard against model extraction
- usage monitoring and anomaly detection for abuse patterns
Tests and signals:
- load and abuse tests (long prompts, recursion, tool loops, extraction attempts)
- spend/latency SLO alerts tied to response playbooks
Evidence to keep: budget policies, monitoring dashboards, incident response records, post-incident fixes.