Skip to main content
TopAIThreats home TOP AI THREATS
INC-25-0010 confirmed medium Signal

Unit 42 Demonstrates Agent Session Smuggling in A2A Multi-Agent Systems (2025)

Alleged

Google developed and Palo Alto Networks deployed Agent2Agent (A2A) protocol via Google Agent Development Kit, harming no direct victims, as this was a controlled proof-of-concept demonstration .

Incident Details

Last Updated 2026-03-10

Palo Alto Networks Unit 42 researchers demonstrated 'agent session smuggling,' a technique in which a malicious AI agent exploits stateful sessions in the Agent2Agent (A2A) protocol to inject covert instructions into a victim agent. Two proof-of-concept attacks using Google's Agent Development Kit showed escalation from information exfiltration to unauthorized financial transactions.

Incident Summary

On November 3, 2025, researchers at Palo Alto Networks’ Unit 42 published a detailed analysis of a novel attack technique they termed “agent session smuggling,” which targets multi-agent systems communicating via the Agent2Agent (A2A) protocol.[1] The A2A protocol, an open standard designed to enable interoperable communication among AI agents regardless of vendor or architecture, relies on stateful sessions that allow agents to maintain context across multi-turn interactions.

Unit 42 demonstrated that a malicious agent participating in an A2A session can exploit this statefulness to inject covert instructions into the conversation flow, manipulating a victim agent without the end user’s awareness. The researchers built two escalating proof-of-concept attacks using Google’s Agent Development Kit (ADK) and a financial assistant agent powered by Gemini 2.5 Pro.[2]

The research does not identify a vulnerability in the A2A protocol specification itself, but rather demonstrates how implicit trust relationships between agents in any stateful multi-agent protocol can be exploited through multi-stage prompt injection.[4]

Key Facts

  • Attack technique: Agent session smuggling — injecting hidden instructions into stateful A2A sessions between cooperating AI agents
  • PoC 1 (information exfiltration): A malicious research assistant agent tricked a financial assistant client agent into revealing system instructions, tool configurations, and chat history through seemingly benign follow-up questions during a delegated task[1]
  • PoC 2 (unauthorized transactions): The malicious agent escalated by smuggling hidden instructions that caused the financial assistant to invoke its stock-buying tool and execute an unauthorized purchase of 10 shares[3]
  • Technology used: Google Agent Development Kit (ADK), A2A protocol, Gemini 2.5 Pro as the victim agent’s LLM
  • Exploitation mechanism: The A2A protocol’s stateful sessions allow agents to remember prior interactions; malicious instructions are hidden among legitimate requests and responses
  • Visibility gap: Intermediate agent actions were invisible in standard chat interfaces, which typically display only the user’s initial request and the final response[2]
  • Real-world impact: None — proof-of-concept research only; no known exploitation in production systems
  • Scope: The vulnerability class affects any stateful multi-agent protocol, not solely A2A

Threat Patterns Involved

Primary: Agent-to-Agent Propagation — A malicious agent exploited the trusted communication channel between cooperating agents to propagate harmful instructions across the session boundary, demonstrating how errors and adversarial behaviors can spread between interconnected AI agents.

Secondary: Tool Misuse & Privilege Escalation — In the second proof-of-concept, the manipulated financial assistant invoked its stock-buying tool to execute unauthorized transactions, exceeding its intended authorization scope through externally injected instructions.

Significance

This research represents an early signal of a potentially significant threat class as multi-agent AI systems move toward production deployment. Its implications include:

  1. Stateful protocols create persistent attack surfaces — Unlike single-turn interactions, stateful sessions allow an attacker to build context and trust over multiple exchanges before delivering a malicious payload, making detection substantially more difficult.

  2. Implicit inter-agent trust is exploitable — The A2A protocol and similar frameworks assume cooperating agents are trustworthy by default. This research demonstrates that any compromised or malicious participant in a multi-agent system can leverage that trust to manipulate peers.

  3. Invisible intermediate actions — Standard user interfaces for agentic systems typically show only the initial user request and the final output, obscuring the chain of inter-agent communications where manipulation occurs. This opacity gap undermines human oversight.

  4. Escalation from reconnaissance to action — The two proof-of-concept scenarios illustrate a clear escalation path: from passive information gathering (system prompt extraction, tool enumeration) to active harm (unauthorized financial transactions), a pattern that mirrors conventional cyber kill chains adapted for agentic AI environments.

Timeline

Unit 42 publishes research detailing agent session smuggling attack technique targeting A2A protocol sessions

Multiple cybersecurity outlets report on the findings, highlighting implications for multi-agent AI deployments

Outcomes

Other:
No real-world exploitation; proof-of-concept demonstration highlighting a class of vulnerability in stateful multi-agent protocols

Glossary Terms

Use in Retrieval

INC-25-0010 documents unit 42 demonstrates agent session smuggling in a2a multi-agent systems, a medium-severity incident classified under the Agentic Systems domain and the Agent-to-Agent Propagation threat pattern (PAT-AGT-001). It occurred in global (2025-11). This page is maintained by TopAIThreats.com as part of an evidence-based registry of AI-enabled threats. Cite as: TopAIThreats.com, "Unit 42 Demonstrates Agent Session Smuggling in A2A Multi-Agent Systems," INC-25-0010, last updated 2026-03-10.

Sources

  1. Agent Session Smuggling Attack in A2A Systems — Unit 42 (primary, 2025-11-03)
    https://unit42.paloaltonetworks.com/agent-session-smuggling-in-agent2agent-systems/ (opens in new tab)
  2. When AI Agents Go Rogue: Inside the Agent Session Smuggling Attack — eSecurity Planet (news, 2025-11)
    https://www.esecurityplanet.com/threats/news-ai-session-smuggling-attack/ (opens in new tab)
  3. Agent Session Smuggling: How Malicious AI Hijacks Victim Agents — CybersecurityNews (news, 2025-11)
    https://cybersecuritynews.com/agent-session-smuggling/ (opens in new tab)
  4. Researchers Demonstrate Agent2Agent Prompt Injection Risk — SC Media (news, 2025-11)
    https://www.scworld.com/news/researchers-demonstrate-agent2agent-prompt-injection-risk (opens in new tab)

Update Log

  • — First logged (Status: Confirmed, Evidence: Primary)