We test and remediate against the most critical risks to large language model applications.
LLM01
Prompt Injection
Direct and indirect prompt injection that manipulates LLM behavior, bypasses guardrails, or extracts sensitive data through crafted inputs.
LLM02
Sensitive Information Disclosure
Data leakage through model outputs—exposing PII, credentials, proprietary information, or training data through context windows and responses.
LLM03
Supply Chain Vulnerabilities
Compromised dependencies, model providers, datasets, or infrastructure that introduce backdoors, biases, or vulnerabilities into deployed AI systems.
LLM04
Data & Model Poisoning
Poisoned training, fine-tuning, or RAG corpora that manipulate LLM behavior—creating backdoors, bias, or enabling targeted manipulation of outputs.
LLM05
Improper Output Handling
Model outputs trusted or executed without validation or policy checks—enabling injection attacks, unauthorized tool invocation, or data corruption in downstream systems.
LLM06
Excessive Agency
Agents granted too much autonomy or permission to act safely—allowing unchecked tool use, privilege escalation, or destructive actions without adequate human oversight.
LLM07
System Prompt Leakage
Extraction of hidden prompts, policies, and tool schemas by attackers—revealing internal instructions and enabling targeted bypass and abuse of safeguards.
LLM08
Vector & Embedding Weaknesses
RAG stores and embeddings that become an attack and data leakage surface—enabling poisoned retrieval, cross-tenant data exposure, and amplified prompt injection.
LLM09
Misinformation
Confident falsehoods, fabricated citations, and misleading outputs that create operational harm, enable fraud, or drive bad decisions without user awareness.
LLM10
Unbounded Consumption
Runaway costs, latency, or capacity exhaustion via abuse or inadequate controls—denial-of-wallet and denial-of-service dynamics targeting AI infrastructure.