Skip to main content
OWASP Top 10 for LLM & Agentic AI

AI & LLM Security
Training

Attackers already use AI to craft phishing emails, clone voices, and hijack LLM assistants. These exercises teach your team to spot the difference.

As organizations adopt AI tools for daily workflows, attackers are weaponizing the same technology. These exercises teach employees to recognize when AI output has been manipulated, when a voice call is synthetic, and when a convincing email was generated by a model.

AI Security Exercises

Hands-on simulations that teach your team to identify and respond to AI-powered attacks. Play available exercises for free, no account required.

1

OWASP Top 10 for LLM Applications

10 exercises · All coming soon

Soon

LLM Prompt Injection Attack

Stop a hidden prompt from hijacking your AI assistant mid-task.

  • Detect hidden instructions embedded in documents processed by AI
  • Trace how injected prompts override legitimate AI behavior
  • Apply safe document handling before feeding content to AI tools
Play Exercise
Soon

Sensitive Data Exposure Through AI

See what happens when confidential data enters a consumer AI tool.

  • Recognize sensitive data categories that should never enter AI prompts
  • Trace how pasted content persists in AI training data and logs
  • Apply data classification policies before using AI tools
Play Exercise
Soon

AI Supply Chain Compromise

Deploy an AI plugin that hides a backdoor in plain sight.

  • Identify supply chain risks in third-party AI models and plugins
  • Detect behavioral anomalies in AI components from external sources
  • Apply vetting procedures before deploying marketplace AI tools
Play Exercise
Soon

AI Training Data Poisoning

Watch poisoned documents corrupt your AI's answers in real time.

  • Trace how manipulated documents alter AI-generated outputs
  • Identify signs of data poisoning in AI responses
  • Apply content integrity controls to knowledge base inputs
Play Exercise
Soon

Unsafe AI Output Handling

Exploit an AI whose outputs flow unchecked into live systems.

  • Identify injection risks when AI outputs feed into downstream systems
  • Trace how unsanitized AI output enables code execution
  • Apply output validation controls between AI and connected systems
Play Exercise
Soon

Over-Permissioned AI Agent

Manipulate an AI assistant into misusing its own permissions.

  • Identify excessive permissions granted to AI agents
  • Trace unauthorized actions performed by manipulated AI tools
  • Apply least-privilege principles to AI agent configurations
Play Exercise
Soon

AI System Prompt Extraction

Extract hidden instructions from a customer-facing AI chatbot.

  • Execute prompt extraction techniques against a live AI chatbot
  • Identify sensitive information exposed through leaked system prompts
  • Apply prompt hardening techniques to prevent system instruction disclosure
Play Exercise
Soon

RAG Pipeline Exploitation

Exploit a RAG pipeline to access documents beyond your clearance.

  • Identify access control failures in vector database retrieval
  • Trace how adversarial embeddings corrupt search results
  • Apply authorization checks at the retrieval layer of RAG systems
Play Exercise
Soon

AI Hallucination and Misinformation

Catch fabricated statistics and fake citations in an AI report.

  • Detect hallucinated facts and fabricated sources in AI outputs
  • Verify AI-generated claims against authoritative references
  • Apply fact-checking workflows to AI-assisted business content
Play Exercise
Soon

AI Denial-of-Service Attack

Launch a denial-of-wallet attack against an unprotected AI API.

  • Identify resource exhaustion vectors in AI API endpoints
  • Trace how crafted prompts escalate compute costs exponentially
  • Apply rate limiting and budget controls to AI service deployments
Play Exercise
2

OWASP Top 10 for Agentic AI Applications

10 exercises · All coming soon

Soon

AI Agent Goal Hijacking

Stop an autonomous AI agent from being redirected by a poisoned email containing hidden instructions.

  • Detect hidden instructions embedded in incoming data that redirect agent objectives
  • Trace how a goal-hijacked agent pivots from legitimate tasks to data exfiltration
  • Apply input validation strategies that prevent agents from treating data as instructions
Play Exercise
Soon

AI Agent Tool Exploitation

Prevent an AI agent from being manipulated into using its legitimate tools to delete files and send unauthorized messages.

  • Identify how ambiguous prompts cause agents to misuse legitimate tool access
  • Trace destructive tool calls triggered by manipulated input parameters
  • Apply least-privilege tool access policies to contain agent tool exploitation
Play Exercise
Soon

Agent Identity and Privilege Abuse

Prevent an AI agent from reusing inherited high-privilege credentials to access systems beyond its authorized scope.

  • Trace how agents inherit and propagate user credentials across different system contexts
  • Identify confused deputy vulnerabilities where agent privilege exceeds intended scope
  • Apply scoped credential delegation to prevent cross-context privilege escalation
Play Exercise
Soon

Agentic AI Supply Chain Attack

Investigate a backdoored third-party AI plugin that silently modifies agent behavior and exfiltrates sensitive data.

  • Detect behavioral anomalies indicating a compromised third-party AI component
  • Trace data exfiltration pathways through backdoored plugins and MCP servers
  • Apply supply chain verification practices before integrating external AI tools
Play Exercise
Soon

AI Agent Code Injection

Catch an AI coding assistant before it executes a shell script containing injected commands that compromise your system.

  • Detect injected commands hidden within AI-generated code and shell scripts
  • Trace how user input flows through code generation into unsandboxed execution
  • Apply code review and sandboxing practices to AI-generated scripts before execution
Play Exercise
Soon

AI Agent Memory Poisoning

Detect adversarial content injected into an AI agent's persistent memory that corrupts all future decisions.

  • Identify poisoned entries in an agent's persistent memory and retrieval context
  • Trace how corrupted memory influences downstream agent decisions across sessions
  • Apply memory integrity verification to detect and remove adversarial content
Play Exercise
Soon

Agent-to-Agent Communication Spoofing

Intercept and identify spoofed messages between AI agents in a multi-agent workflow before fabricated instructions cause damage.

  • Detect spoofed agent identities and fabricated messages in multi-agent communication channels
  • Trace how unauthenticated inter-agent messages enable man-in-the-middle attacks
  • Apply message authentication and agent identity verification to secure multi-agent systems
Play Exercise
Soon

Multi-Agent Cascading Failure

Contain a minor AI hallucination before it cascades through downstream agents into a catastrophic system-wide failure.

  • Trace error propagation from a single agent hallucination through multiple downstream systems
  • Identify amplification points where small errors compound into catastrophic outcomes
  • Apply circuit breaker patterns and human checkpoints to interrupt cascading agent failures
Play Exercise
Soon

Over-Trusting AI Agent Recommendations

Catch a series of compromised AI agent recommendations that exploit your trust to approve a fraudulent transfer and a backdoored code change.

  • Recognize automation bias patterns where consistent AI accuracy creates false confidence
  • Identify subtle anomalies in AI recommendations that indicate manipulation or compromise
  • Apply structured verification workflows that resist social engineering through AI interfaces
Play Exercise
Soon

Detecting a Rogue AI Agent

Investigate a compromised AI agent that appears functional while silently performing unauthorized actions and evading monitoring.

  • Detect covert unauthorized actions performed by an agent that appears to be operating normally
  • Trace persistence mechanisms that allow rogue agents to survive restarts and monitoring sweeps
  • Apply behavioral analysis and anomaly detection to distinguish rogue agents from legitimate ones
Play Exercise

What Is AI Security Training?

AI security training prepares employees to recognize and respond to threats that exploit artificial intelligence, large language models, and autonomous AI agents. As organizations integrate AI assistants into document analysis, code review, customer support, and decision-making workflows, attackers target these same tools to steal data, manipulate outputs, and bypass security controls.

This catalogue covers 21 exercises organized into two OWASP-aligned courses. The OWASP Top 10 for LLM Applications covers prompt injection, sensitive data exposure, supply chain compromise, data poisoning, unsafe output handling, excessive agency, system prompt leakage, RAG pipeline exploitation, AI misinformation, and denial-of-service attacks. The OWASP Top 10 for Agentic AI Applications covers goal hijacking, tool exploitation, identity and privilege abuse, agentic supply chain attacks, code injection, memory poisoning, inter-agent communication spoofing, cascading failures, trust exploitation, and rogue agents.

These exercises use interactive 3D simulations where employees practice identifying manipulated AI output, compromised agents, and AI-powered attacks in realistic workplace scenarios.

Frequently Asked Questions

Common questions about AI security threats and how training helps defend against them.

What is AI prompt injection?

AI prompt injection is an attack where malicious instructions are hidden inside documents, emails, or web pages that an AI assistant processes. When the AI reads the content, it follows the hidden instructions instead of the user's intent.

This can cause the AI to leak sensitive data, ignore safety rules, or perform unauthorized actions without the user realizing the input was manipulated.

How can prompt injection lead to data exfiltration?

An attacker embeds instructions in a document telling the AI to include sensitive data in its output, encode it in URLs, or send it to external endpoints.

Because the AI processes the document's full text, it may follow these instructions alongside legitimate content, sending confidential information to unintended recipients.

Why is AI security training important for employees?

As organizations integrate AI tools into daily workflows, employees interact with LLMs for document analysis, code review, customer support, and decision-making.

Without proper training, staff cannot recognize when AI-generated output has been manipulated, when a deepfake voice call is impersonating a colleague, or when an AI-powered phishing email bypasses traditional detection methods. AI security training closes this gap before attackers exploit it.

What are the biggest AI-related security threats?

The most pressing threats include prompt injection attacks that hijack AI assistants, deepfake voice and video used for impersonation and fraud, AI-generated phishing emails that are nearly indistinguishable from legitimate messages, and chatbot manipulation that extracts sensitive data from enterprise AI systems.

These threats are growing as AI adoption accelerates across industries.

What is the OWASP Top 10 for LLM Applications?

The OWASP Top 10 for LLM Applications is an industry-standard framework that identifies the ten most critical security risks in large language model deployments.

It covers prompt injection (LLM01), sensitive information disclosure (LLM02), supply chain vulnerabilities (LLM03), data poisoning (LLM04), improper output handling (LLM05), excessive agency (LLM06), system prompt leakage (LLM07), vector and embedding weaknesses (LLM08), misinformation (LLM09), and unbounded consumption (LLM10). Our course includes one hands-on exercise for each risk.

What is the OWASP Top 10 for Agentic AI Applications?

The OWASP Top 10 for Agentic AI Applications is a 2025 framework that addresses security risks specific to autonomous AI agents that use tools, make decisions, and take actions independently.

It covers agent goal hijacking (ASI01), tool misuse and exploitation (ASI02), identity and privilege abuse (ASI03), agentic supply chain vulnerabilities (ASI04), unexpected code execution (ASI05), memory and context poisoning (ASI06), insecure inter-agent communication (ASI07), cascading failures (ASI08), human-agent trust exploitation (ASI09), and rogue agents (ASI10).

Start Training Your Team on AI Threats

Start with free interactive exercises or request a demo to see how RansomLeak's AI security training fits your organization.