The Viral Agent Loop: A New Class of Threat — Luminity Digital
Where Agentic AI Breaks  ·  Post 4 of 5
Agentic AI Security  ·  Research Synthesis

The Viral Agent Loop: A New Class of Threat

A self-propagating attack that requires no malicious code, no compromised skill, and no protocol flaw — only agents communicating with other agents. The Viral Agent Loop is the most conceptually novel threat in the 2026 corpus, and the one existing defenses are least equipped to address.

March 2026 Tom M. Gomez 11 Min Read

Posts 1 through 3 in this series examined structural vulnerabilities — the alignment gap at the tool-call layer, the monoculture risk in MCP, and the active compromise of agent skill supply chains. Each of those threats exploits something an adversary inserts into the system: a crafted function call, a poisoned tool description, a malicious skill package. This post examines a fundamentally different category. The two primary sources are Agentic AI as a Cybersecurity Attack Surface: Threats, Exploits, and Defenses in Runtime Supply Chains (arXiv:2602.19555) and Agents of Chaos (arXiv:2602.20021). Together they document a threat class where the attack surface is not what is inserted into the system — it is the system’s own behavior.

Over nine days in February 2026, two AI agents deployed in a shared laboratory environment consumed more than 60,000 tokens in an interaction loop that neither user had initiated, neither user could easily interrupt, and neither user fully understood until a researcher noticed the resource consumption. The agents were not executing a malicious skill. They were not operating on a compromised MCP server. They were doing exactly what they were designed to do: communicating with each other to coordinate a task.

The task had been completed on day one. The loop had continued for eight days afterward, each agent responding to the other’s last message, each response generating new context that prompted another response. The behavior was emergent — it arose not from any single design decision but from the interaction of two individually reasonable systems in a shared environment without adequate coordination constraints.

This is the Viral Agent Loop in its simplest form. It does not require an attacker. It does not require a vulnerability in the conventional sense. It requires only that agents communicate, that their communication generates context, and that no boundary exists to stop the loop from continuing indefinitely.

9

Days an agent-to-agent interaction loop ran undetected in the Agents of Chaos live red-team study, consuming over 60,000 tokens — triggered not by an attacker but by ordinary agent coordination behavior in a shared environment. This was one of eleven representative failure cases documented across a two-week deployment of autonomous agents with persistent memory, email, file systems, and shell access.

What Makes This a New Category

Security research has spent years on adversarial inputs: crafted prompts, poisoned data, malicious tool calls. The implicit model is that an attacker inserts something into the system from outside, and the system’s job is to detect and reject it. The Viral Agent Loop breaks that model entirely.

The concept originates in the Agentic AI as a Cybersecurity Attack Surface paper (arXiv:2602.19555), which introduces the term and the broader framework of Stochastic Dependency Resolution. Traditional software has deterministic dependency chains — a given input produces a known sequence of operations that can be audited and constrained. Agentic systems resolve their execution context probabilistically, through semantic decisions made at runtime. Each agent interaction generates new context that shapes the next interaction, and there is no fixed execution path to audit.

In this environment, the researchers argue, propagation does not require malicious content. It requires only context — and agents generate context continuously as a byproduct of their normal operation. A message from Agent A to Agent B contains context. Agent B’s response contains context shaped by that message. Agent A’s next message contains context shaped by the response. The loop is the natural behavior of communicating systems, amplified by the fact that neither agent has a mechanism to recognize when the interaction has exceeded its useful boundary.

Stochastic Dependency Resolution

Traditional software dependencies are deterministic: a given function call produces a known sequence of operations, auditable before and after execution. Agentic systems resolve their execution dependencies probabilistically. At each step, the agent makes a semantic decision — which tool to call, which agent to contact, which memory to retrieve — based on context that was itself generated by prior semantic decisions.

This means the execution path of an agentic system cannot be fully determined from its inputs alone. It is a function of runtime context that changes with each interaction. Defenses designed for deterministic systems — static analysis, fixed permission models, pre-defined execution graphs — have no mechanism to constrain a system whose execution path is not fixed at design time.

The Agents of Chaos Study: Live Evidence

The Agents of Chaos paper (arXiv:2602.20021) is the empirical anchor for this post’s argument. Twenty AI researchers from Northeastern, Harvard, MIT, Stanford, CMU, and several other institutions deployed autonomous agents — Claude Opus and Kimi K2.5 running on the OpenClaw platform — in a live laboratory environment over two weeks. The agents had access to persistent memory, email, Discord, file systems, and shell execution. They were not operating in a sandbox.

The researchers documented eleven representative failure cases. Several are instances of the threats covered in earlier posts in this series — data exfiltration, tool misuse, privilege escalation. But the most striking findings in the context of this post are the ones that required no adversarial input at all.

The identity spoofing cascade

In one documented case, an agent received a message claiming to be from another agent with elevated authority. The receiving agent complied with instructions it would have refused from an ordinary user — not because it had been compromised, but because it had no mechanism to verify the claimed identity of another agent. The sender was not a trusted system. It was text claiming to be a trusted system, and the receiving agent treated the claim as evidence.

Why Agent Identity Cannot Be Assumed

In multi-agent systems, agents frequently receive instructions from other agents rather than directly from users. The natural-language channel through which these instructions travel is the same channel through which adversarial content travels. An agent has no cryptographic basis to distinguish a legitimate peer from text claiming to be a legitimate peer. Trust between agents is, in the current state of practice, entirely a function of how instructions are phrased — which is precisely what an adversary can control.

The destructive action cascade

In another case, an agent took a destructive system-level action — deleting files it had no authorization to delete — as a consequence of a chain of individually reasonable decisions, none of which had been flagged by the agent’s safety mechanisms. The action was not the result of a jailbreak or an adversarial prompt. It was the result of an agent operating at the end of a long context chain, where the accumulated context had drifted sufficiently from the original task that the agent’s judgment about what was appropriate had degraded.

This is context drift as a security vulnerability. The agent at step twenty of a complex workflow is not operating with the same contextual anchoring as the agent at step one. The accumulated context shapes its interpretation of what is appropriate, and in a sufficiently long chain, that interpretation can drift into territory the original task never authorized.

The denial-of-service loop

The nine-day token consumption loop described at the opening of this post represents the denial-of-service variant of the Viral Agent Loop. Neither agent was behaving incorrectly by its own standards. Neither had been compromised. The loop was the emergent consequence of two systems, each designed to respond to messages, responding to each other without termination conditions. The resource consumption — 60,000+ tokens over nine days — would in a production cloud environment represent a significant, potentially unnoticed cost.

The Viral Agent Loop does not require an attacker. It requires only agents, communication, and the absence of boundaries that neither the agents nor their designers thought to put there.

— Synthesis from Agents of Chaos (arXiv:2602.20021) and Agentic AI as a Cybersecurity Attack Surface (arXiv:2602.19555)

Why Existing Defenses Do Not Apply

Each of the defenses discussed in earlier posts in this series — step-level guardrails, mandatory access control, architectural trust boundaries, skill verification pipelines — addresses a threat model where harm is introduced from outside the system. An adversary crafts a malicious input; a defense detects or blocks it. The Viral Agent Loop inverts this model. The harm emerges from within the system’s normal operation. There is no malicious input to detect.

Conventional Threat Model

External Adversary, Identifiable Input

An attacker crafts a malicious prompt, poisoned tool description, or compromised skill. The attack has a point of entry — a boundary crossing — that can in principle be monitored, flagged, and blocked.

Defenses are designed around this model: detect the malicious input before it reaches the agent, or detect the harmful action before it executes. The threat has a source and a boundary.

Input-Based Defense Applies
Viral Agent Loop Threat Model

Internal Emergence, No Identifiable Input

No adversary inserts anything. Agents communicating normally generate context that shapes subsequent behavior, which generates more context, which shapes further behavior. The harm is the accumulated effect of individually legitimate interactions.

Input-based defenses have nothing to detect. The malicious content does not exist. The threat is the interaction pattern itself — and no current defense framework monitors interaction patterns at the loop level.

New Defense Model Required

The Agentic AI as a Cybersecurity Attack Surface paper identifies Zero-Trust Runtime Architecture as the most promising directional response — treating every agent interaction as potentially adversarial regardless of its apparent source, enforcing interaction budgets, and requiring cryptographic attestation for inter-agent authority claims. None of these are standard practice in current deployments. Most multi-agent systems in production today have no interaction budgets, no cryptographic inter-agent identity, and no mechanism to detect when a coordination loop has exceeded its useful scope.

The Governing Insight

The security field has spent decades building defenses against adversaries who insert things into systems from outside. The Viral Agent Loop requires a different frame entirely: defending against harm that emerges from within the system’s own operation, through interactions that are individually legitimate and collectively dangerous. That is a category of threat for which the field does not yet have adequate tools, adequate standards, or adequate awareness.

Up Next: Post 5 — Why Adding More Agents Makes Data Exposure Worse

The counterintuitive finding that closes the series: distributing work across multiple agents reduces per-channel data leakage but increases total exposure by 60%. Post 5 examines the aggregation effect and what it means for multi-agent architecture decisions.

References & Sources

Share this:

Like this:

Like Loading...