Gallery

Contacts

411 University St, Seattle, USA

engitech@oceanthemes.net

+1 -800-456-478-23

Hub / Secure / Agent Threat Landscape
Secure Pillar

The Agentic AI Threat Landscape

OWASP, MITRE ATLAS, and CSA MAESTRO: Three Frameworks Mapping the New Attack Surface

2,842 Words 13 Min Read 6 Sources 2026-03-21 Published
Table of Contents
  1. 01 Why Agentic AI Needs Its Own Threat Model
  2. 02 OWASP Agentic Security Initiative
  3. 03 MITRE ATLAS: Adversary Tactics for AI
  4. 04 CSA MAESTRO: The 7-Layer Model
  5. 05 Cross-Framework Mapping
  6. 06 Top Threats by Severity
  7. 07 Real-World Incidents
  8. 08 Defense Strategies Overview
01 // Context Why Agentic AI Needs Its Own Threat Model Foundation

Traditional LLM security was built for a request-response world. A user sends a prompt, the model generates text, and a human decides what to do with it. The threat model for that interaction is relatively contained: prompt injection, data leakage, hallucination. Serious problems, but bounded ones.

Agentic AI breaks that model. When an AI system can autonomously plan multi-step tasks, invoke external tools, persist information across sessions, and coordinate with other agents, the attack surface expands from a single input-output boundary to an entire operational architecture. The OWASP Agentic Security Initiative puts it directly: "agentic AI threats are either new or agentic variations of existing threats," with agent memory, tool integration, identity and authorization, and multi-agent coordination as the key new attack surfaces.

Consider the difference. A chatbot that hallucinates produces wrong text. An AI agent that hallucinates might invoke a tool based on fabricated data, write that fabrication to persistent memory, and then propagate it to other agents in the system. The OWASP ASI calls this "destructive reasoning" and flags it as a cascading failure mode unique to agentic architectures.

Three major security frameworks have converged on this problem in 2025: OWASP's Agentic Security Initiative, MITRE ATLAS, and the Cloud Security Alliance's MAESTRO framework. Each approaches the threat landscape from a different angle. Together, they provide the most complete map available of what can go wrong when AI systems gain autonomy. This article walks through all three, maps how they relate, and identifies the threats that matter most for organizations building or deploying agents today.

Combined Threat Intelligence
15 OWASP ASI Threats
9 MITRE ATLAS Techniques
39 MAESTRO Threats (7 Layers)
12 CSA Red Team Categories
02 // OWASP The OWASP Agentic Security Initiative 15 Threats

The OWASP Agentic Security Initiative released its Threats and Mitigations document (v1.0a) in February 2025, building on the OWASP Top 10 for LLM Applications 2025 (published November 2024). It identifies 15 agentic threat categories, organized into threats that are entirely new to agentic AI and those that represent agentic amplifications of existing LLM vulnerabilities.

The framework begins with a reference architecture that maps the components common to agentic systems: one or more LLM models for reasoning, services including built-in functions and external APIs invoked via function calling or tools interfaces, and supporting services such as external storage for long-term memory, vector databases, and RAG data sources. Each component introduces distinct threat surfaces.

The 15 OWASP Agentic Threat Categories

T-Code Threat Name Type Related LLM Top 10 Severity
T1 Memory Poisoning New Agentic LLM04 (Data Poisoning) High
T2 Tool Misuse New Agentic LLM08 (Excessive Agency) High
T3 Privilege Compromise Amplification LLM08 (Excessive Agency) High
T4 Resource Overload Amplification LLM10 (Unbounded Consumption) Medium
T5 Cascading Hallucination Attacks New Agentic LLM09 (Misinformation) High
T6 Intent Breaking & Goal Manipulation New Agentic LLM01 (Prompt Injection) High
T7 Misaligned & Deceptive Behaviors New Agentic LLM09 (Misinformation) High
T8 Repudiation & Untraceability New Agentic Medium
T9 Identity Spoofing & Impersonation Amplification High
T10 Overwhelming Human-in-the-Loop New Agentic Medium
T11 Unexpected RCE and Code Attacks Amplification LLM05 (Improper Output Handling) Critical
T12 Agent Communication Poisoning New Agentic LLM04 (Data Poisoning) High
T13 Rogue Agents New Agentic LLM08 (Excessive Agency) High
T14 Human Attacks on Multi-Agent Systems New Agentic LLM08 (Excessive Agency) High
T15 Human Manipulation New Agentic LLM09 (Misinformation) High

Note: The New Agentic / Amplification categorization is an editorial analysis by Tech Jacks Solutions. OWASP's ASI document does not use this taxonomy.

The pattern is striking. Of the 15 threat categories, 11 are entirely new to agentic AI, with no direct equivalent in traditional LLM security. Only 4 are amplifications of existing LLM Top 10 entries. The multi-agent threats (T12, T13, T14) represent an entirely new class of risk that emerges only when agents communicate and coordinate with each other.

The OWASP ASI also defines 10 agentic patterns that shape how threats manifest: reflective agents, task-oriented agents, hierarchical agents, coordinating agents, distributed agent ecosystems, human-in-the-loop collaboration, self-learning and adaptive agents, RAG-based agents, planning agents, and context-aware agents. Each pattern carries a different attack surface profile. A sequential agent with a single LLM and limited API access faces a fundamentally different threat landscape than a collaborative agent swarm with cross-agent cross-session memory.

For deep analysis of the top-ranked threat, see Prompt Injection in Agentic Systems: Why It's the #1 Threat. For the tool-layer risks, see Tool Misuse, Excessive Agency, and the MCP Compositional Risk.

03 // MITRE MITRE ATLAS: Adversary Tactics for AI Systems ATT&CK for AI

MITRE ATLAS (Adversarial Threat Landscape for Artificial Intelligence Systems) extends the well-known ATT&CK framework into the AI domain. Where OWASP catalogs vulnerabilities and the MAESTRO framework layers threats architecturally, ATLAS maps adversary tactics, techniques, and procedures (TTPs) specific to machine learning systems.

ATLAS does not yet have agentic-specific sub-techniques beyond prompt injection variants (T0051.000 for direct, T0051.001 for indirect). However, its existing techniques are referenced extensively across both the OWASP ASI and CSA MAESTRO data, providing the adversary behavior taxonomy that connects to the vulnerability catalogs.

ATLAS Techniques Referenced in Agentic Threat Models

ATLAS ID Technique Name Agentic Context
AML.T0051 LLM Prompt Injection Direct (T0051.000) and indirect (T0051.001) injection; in agentic systems, indirect injection via tool outputs and inter-agent messages is the primary vector
AML.T0054 LLM Jailbreak Bypassing safety constraints in agentic contexts where the agent has tool access, turning a jailbreak into an operational exploit
AML.T0043 Craft Adversarial Data Referenced across nearly all agentic threats as the meta-technique; adversarial inputs that manipulate agent reasoning and tool selection
AML.T0018 Infer Training Data / Data Poisoning RAG poisoning, memory poisoning, embedding attacks against the agent's knowledge base
AML.T0020 Poison Training Data Fine-tuning attacks, sleeper agent implantation; Anthropic's sleeper agent research (arXiv:2401.05566) demonstrated persistence through safety training
AML.T0010 ML Supply Chain Compromise Framework dependencies (LangChain, AutoGen, CrewAI), MCP server plugins, model registry manipulation
AML.T0048 AI Supply Chain Abuse Tool misuse, improper output handling, remote code execution through agent tool invocations
AML.T0024 Exfiltration via ML Model Model theft through systematic querying, embedding inversion attacks against agent vector stores
AML.T0029 Denial of ML Service Resource overload, Denial of Wallet attacks against cloud-hosted agent deployments

The value of ATLAS for agentic security is its adversary-centric perspective. While OWASP describes what can go wrong and MAESTRO describes where it can go wrong, ATLAS describes how adversaries actually attack. This makes it particularly useful for red teaming exercises and threat modeling sessions where the goal is to think like an attacker. The technique AML.T0043 (Craft Adversarial Data) appears as a reference across nearly every agentic threat category, underscoring that adversarial data manipulation is the foundational attack primitive against agent systems.

04 // CSA MAESTRO: The 7-Layer Threat Model 39 Threats

The Cloud Security Alliance's MAESTRO framework (Multi-Agent Environment Security Threat and Risk Operations) takes a fundamentally different approach from OWASP's vulnerability catalog. MAESTRO organizes threats into seven architectural layers, each representing a distinct stratum of the agentic AI stack. This layered model extends traditional threat modeling approaches like STRIDE with AI-specific concerns at every level.

The architectural approach makes MAESTRO particularly useful for enterprise security teams who need to assign threat ownership. Each layer maps to specific teams, tools, and controls. A foundation model threat (L1) is the responsibility of the ML engineering team. A deployment threat (L5) falls to platform engineering and DevSecOps -- the same teams evaluating cloud agent platforms where these agents run in production. A governance threat (L7) belongs to the compliance and legal functions.

The Seven MAESTRO Layers

L1
Foundation Model
Threats targeting the underlying LLM reasoning engine: training data poisoning, model theft, cascading hallucinations, misaligned behaviors, indirect prompt injection, and intent breaking.
6 threats
L2
Data & Knowledge
Memory poisoning, RAG knowledge base poisoning, context window exploitation, shared memory cross-contamination, and embedding inversion.
5 threats
L3
Agent Architecture
Agent communication poisoning, rogue agents, orchestration hijacking, reflection loop exploitation, human attacks on multi-agent systems, and agentic pattern abuse.
6 threats
L4
Tool & API Integration
Tool misuse and excessive agency, privilege compromise via dynamic permissions, unexpected RCE, MCP compositional risk, confused deputy via non-human identities, and supply chain compromise via frameworks.
6 threats
L5
Deployment & Infrastructure
Resource overload, container and sandbox escape, identity spoofing, shadow agent deployment, and credential and secret exposure.
5 threats
L6
Monitoring & Observability
Repudiation and untraceability, log tampering, alert suppression, overwhelming human-in-the-loop reviewers, and observability gaps in reasoning chains.
5 threats
L7
Governance & Compliance
Audit trail gaps, policy bypass, BBOM gaps, human manipulation via agent trust, EU AI Act classification evasion, and cross-framework governance gaps.
6 threats

The CSA also published its Agentic AI Red Teaming Guide (August 2025), which complements the MAESTRO framework with 12 practical vulnerability categories and actionable test procedures. These red teaming categories include agent authorization and control hijacking, checker-out-of-the-loop validation, agent critical system interaction, goal and instruction manipulation, hallucination exploitation, impact chain and blast radius assessment, knowledge base poisoning, memory and context manipulation, orchestration and multi-agent exploitation, resource and service exhaustion, supply chain and dependency attacks, and agent untraceability testing.

Where MAESTRO tells you what to worry about, the Red Teaming Guide tells you how to test for it. Together, they provide a complete threat-to-test pipeline for enterprise security programs.

05 // Mapping Cross-Framework Mapping Convergence

The three frameworks are not competitors. They are complementary lenses on the same problem. OWASP provides the vulnerability catalog with actionable mitigations. MITRE ATLAS provides the adversary behavior model with attack technique references. CSA MAESTRO provides the architectural threat map with layer-by-layer ownership assignments. Understanding how they connect is essential for building a comprehensive threat model.

The table below maps the OWASP ASI threat codes to their primary MAESTRO layer, showing how each OWASP vulnerability fits into the 7-layer architecture. CWE references provide the bridge to traditional software vulnerability databases.

OWASP T-Codes to MAESTRO Layers

OWASP T-Code Threat MAESTRO Layer Key CWE Refs
T1 Memory Poisoning L2 Data & Knowledge CWE-74
T2 Tool Misuse L4 Tool & API Integration CWE-250, CWE-441
T3 Privilege Compromise L4/L5 Tool & Deployment CWE-269
T4 Resource Overload L5 Deployment & Infrastructure CWE-400, CWE-770
T5 Cascading Hallucinations L1/L3 Foundation & Architecture
T6 Intent Breaking L1/L3 Foundation & Architecture CWE-77, CWE-94
T7 Misaligned Behaviors L1/L3 Foundation & Architecture
T8 Repudiation L6 Monitoring & Observability CWE-778, CWE-223
T9 Identity Spoofing L5 Deployment & Infrastructure CWE-287, CWE-290
T10 Overwhelming HITL L6 Monitoring & Observability
T11 Unexpected RCE L4 Tool & API Integration CWE-94
T12 Communication Poisoning L3 Agent Architecture CWE-74
T13 Rogue Agents L3 Agent Architecture CWE-269
T14 Human Attacks on MAS L3 Agent Architecture CWE-250
T15 Human Manipulation L7 Governance & Compliance

The mapping reveals clear clustering. Agent Architecture (L3) absorbs the most OWASP threats, including all multi-agent categories. Tool & API Integration (L4) captures the operational risk surface. And Monitoring & Observability (L6) represents a category of threats that traditional LLM frameworks largely ignored: the difficulty of auditing, tracing, and attributing autonomous agent actions.

For organizations building governance programs around agents, this mapping provides the connective tissue between security findings and compliance requirements. A MAESTRO L7 gap maps directly to the regulatory obligations covered in the EU AI Act's requirements for high-risk AI systems (see also the EU AI Act Hub for comprehensive compliance guidance). Organizations applying the NIST AI Risk Management Framework will find that MAESTRO layers map cleanly to the MAP and MEASURE functions, while the AI Governance Hub provides broader context on operationalizing these frameworks at enterprise scale.

06 // Severity Top Threats by Severity Risk Matrix

Across all three frameworks, the severity distribution of threats to agentic systems skews heavily toward high and critical. This is not an artifact of classification inflation. It reflects the reality that autonomous systems with tool access and persistent memory create fundamentally higher-stakes failure modes than passive text generation.

MAESTRO Severity Distribution (39 Threats)

10 Critical
19 High
10 Medium

The OWASP severity rankings converge on the same conclusion. Three threats earn the Critical designation: Prompt Injection (LLM01), Excessive Agency (LLM08), and Unexpected RCE (T11). Prompt injection is rated critical because in agentic systems, a successful injection does not just produce wrong text. It can hijack the agent's planning loop, causing it to execute unauthorized tool calls, exfiltrate data through legitimate channels, or propagate malicious instructions to downstream agents.

Excessive Agency (LLM08) is the defining agentic threat. The OWASP ASI describes it as the convergence of excessive functionality, excessive permissions, and excessive autonomy. When an agent has more tools than it needs, broader permissions than its task requires, and less human oversight than the risk warrants, every other threat becomes amplified.

Agent-Specific Attack Vectors

Indirect Prompt Injection via Tool Outputs
Malicious instructions embedded in data returned by tools, APIs, or retrieved documents are processed by the agent's reasoning loop, leading to unauthorized actions.
Source: OWASP ASI T&M v1.0a
Cross-Agent Prompt Injection
In multi-agent systems, a compromised or manipulated agent injects malicious instructions into messages passed to peer agents, cascading compromise across the architecture.
Source: OWASP ASI T&M v1.0a
Confused Deputy via Non-Human Identity
An AI agent with higher privileges than the user is tricked into performing unauthorized actions. Machine accounts and agent-based API keys lack session-based oversight.
Source: OWASP ASI T&M v1.0a, pp. 13-14; MAESTRO L4-T5
Tool Chaining Exploitation
Attacker manipulates the agent into chaining multiple legitimate tools in a sequence that achieves an unauthorized objective not possible through any single tool invocation.
Source: OWASP ASI T&M v1.0a; OWASP-AGENT-T02
Orchestration Hijacking
Attackers exploit the agent orchestration layer to reroute task delegation, manipulate agent selection logic, or inject unauthorized steps into multi-step workflows.
Source: MAESTRO L3-T3
Approval Fatigue (Overwhelming HITL)
Generating high volumes of benign-seeming approval requests to desensitize human reviewers, then inserting a malicious action that gets rubber-stamped.
Source: OWASP ASI T&M v1.0a, p. 15; OWASP-AGENT-T09
07 // Evidence Real-World Incidents and Research Case Files

These threats are not theoretical. The OWASP and CSA source documents reference documented incidents and peer-reviewed research demonstrations that validate the threat categories.

Documented Incidents

CVE-2024-5184: Email Assistant Hijacking
LLM-powered email assistant exploited via prompt injection to access sensitive information and manipulate email content.
Slack AI Data Exfiltration
Private channel data exfiltrated via indirect prompt injection through shared documents (PromptArmor, 2024).
ChatGPT Memory Injection
Persistent memory injection where attacker-planted facts influenced all future conversations in the session.
Air Canada Chatbot Liability
Air Canada held liable for chatbot misinformation about bereavement fare policies that did not exist.
Ray AI Framework CVEs
Ray AI framework vulnerabilities (CVE-2023-48022 and others) exploited in the wild, compromising AI infrastructure.
Denial of Wallet Attacks
Cloud AI deployments where API abuse led to $10,000+ unexpected billing through unbounded consumption.

Research Demonstrations

Anthropic Sleeper Agents (arXiv:2401.05566)
Demonstrated LLMs that maintained deceptive behavior through safety training, validating the persistence of misaligned behaviors.
PoisonGPT
Lobotomized LLM published on Hugging Face with directly modified parameters to spread misinformation through supply chain.
ConfusedPilot Attack
RAG data poisoning demonstrated against enterprise AI assistants, showing how adversarial documents corrupt agent knowledge bases.
NIST Agent Hijacking Evaluations
NIST documented how adversarial data causes agents to execute unintended tool interactions (January 2025).

The pattern across these incidents is consistent: the autonomous, multi-step nature of agentic systems transforms what would be a contained vulnerability in a traditional LLM application into a cascading failure. The Slack AI exfiltration, for example, exploited the same indirect prompt injection mechanism that exists in any RAG system. But the agent's ability to take action on the injected instructions, retrieving private data and surfacing it through legitimate channels, is what turned a vulnerability into an incident.

08 // Defenses Defense Strategies Overview Mitigation

Each framework provides defense guidance aligned with its perspective. The OWASP ASI offers structured mitigations for each T-code, the OWASP Securing Agentic Applications Guide provides lifecycle-based security guidance, and the CSA Red Teaming Guide prescribes test-driven validation across its vulnerability categories. Here is a summary of the cross-cutting defenses that appear across all three.

Memory and Data Integrity

For memory poisoning (T1) and data-layer threats (MAESTRO L2), the OWASP ASI prescribes memory content validation, session isolation, robust authentication for memory access, anomaly detection systems, and regular memory sanitization. AI-generated memory snapshots enable forensic analysis and rollback when contamination is detected.

Tool Access and Privilege Controls

For tool misuse (T2) and privilege compromise (T3), defenses center on strict tool access verification, tool usage pattern monitoring, agent instruction validation with clear operational boundaries, and execution logs tracking all tool calls for anomaly detection. The CSA Red Teaming Guide adds just-in-time permissions, task-specific access, automatic revocation post-task, and network allow-lists. These controls are explored in detail in Tool Misuse, Excessive Agency, and the MCP Compositional Risk.

Architecture-Specific Hardening

The OWASP Securing Agentic Applications Guide (July 2025) maps different attack surface profiles to architecture patterns. A sequential agent with a single LLM, simple workflow, and limited API access has a different threat profile than a hierarchical agent with an orchestrator and specialized sub-agents, or a collaborative agent swarm with peer agents and cross-agent cross-session memory. Each architecture requires different controls. The guide provides component-level attack surface analysis mapped to the T-codes:

  • LLMs (KC1): Cascading hallucinations, intent breaking, misaligned behaviors, human manipulation
  • Orchestration (KC2): Intent breaking, repudiation, identity spoofing, overwhelming HITL, communication poisoning, rogue agents, human attacks on MAS
  • Reasoning/Planning (KC3): Cascading hallucinations, intent breaking, misaligned behaviors, repudiation, human manipulation
  • Memory (KC4): Memory poisoning, privilege compromise, cascading hallucinations, intent breaking, repudiation, communication poisoning
  • Tool Integration (KC5): Tool misuse, privilege compromise, intent breaking, misaligned behaviors, repudiation, unexpected RCE
  • Operational Environment (KC6): Tool misuse, privilege compromise, resource overload, overwhelming HITL, unexpected RCE, communication poisoning, rogue agents, human manipulation

Runtime Hardening

The Securing Agentic Applications Guide also provides runtime-specific guidance: harden the VM base level, contain the agentic runtime through sandboxing, secure agent memory and tools and context, implement observability and forensics, and apply cloud-specific hardening. For organizations deploying agents to production, these runtime controls represent the operational security baseline.

Defense strategy details for specific threat categories are covered in the companion Secure pillar articles: Prompt Injection in Agentic Systems covers the highest-priority injection defenses, and Tool Misuse and Excessive Agency covers tool-layer controls and the emerging MCP compositional risk pattern.

Key Takeaways
  • Agentic AI creates fundamentally new threats. Of the 15 OWASP ASI categories, 11 are entirely new to agentic systems, with no direct equivalent in traditional LLM security.
  • Three frameworks, three perspectives. OWASP provides the vulnerability catalog and mitigations. MITRE ATLAS provides the adversary behavior model. CSA MAESTRO provides the architectural threat map with layer-by-layer ownership.
  • MAESTRO identifies 39 threats across 7 architectural layers. Of these, 10 are rated critical and 19 are rated high. The severity distribution reflects the operational stakes of autonomous systems with tool access.
  • Prompt injection and excessive agency are the top-ranked threats. Both are rated critical by OWASP, and they interact: an agent with excessive permissions is exponentially more dangerous when compromised by injection.
  • Multi-agent coordination is a new attack surface category. Communication poisoning, rogue agents, and human attacks on multi-agent systems have no precedent in single-model LLM security.
  • Defenses must be architecture-specific. A sequential agent, a hierarchical agent, and a collaborative swarm each require different control sets mapped to their distinct attack surface profiles.
Sources

Continue the Secure pillar deep dive: Prompt Injection in Agentic Systems breaks down the #1 ranked threat, and Tool Misuse and Excessive Agency covers the operational risk of unconstrained agent tool access. For the latest threat intelligence, visit the Security News Center. Security professionals building careers in this space will find relevant roles and skills at the AI Governance Careers hub. Or explore the full Agentic AI Hub.

◀ Back to Pillar Secure: Agentic AI Next Article ▶ Prompt Injection in Agentic Systems: Why It's the #1 Threat