Capability Restriction Is Not Security: The Structural Controls Agentic AI Actually Needs — Luminity Digital
Agentic AI Security

Capability Restriction Is Not Security: The Structural Controls Agentic AI Actually Needs

Organizations deploying AI agents are making a foundational security error: they are treating the reduction of what an agent can do as a substitute for controlling what an agent is authorized to do. The first is hygiene. The second is security. Mistaking one for the other leaves production agentic systems architecturally vulnerable in ways that no amount of tool-list trimming can address.

March 2026
Tom M. Gomez
14 Min Read

The distinction between capability restriction and structural security controls is not one of degree — it is one of kind. Capability restriction asks: what can we take away from this agent so it has less opportunity to cause harm? Structural security controls ask: regardless of what this agent is capable of, what is it actually authorized to do right now, under this verified identity, in this specific context, and how do we enforce that at a layer the agent cannot reason its way around? The first question produces a smaller attack surface. The second question produces a security posture. Enterprise agentic AI deployments require both — but the industry is currently building the first and calling it the second.

When an enterprise security team reviews an AI agent deployment and concludes that the agent is “secured” because it has been given read-only database tokens, stripped of email-sending tools, and limited to a curated set of APIs, they have performed a meaningful risk-reduction exercise. The agent genuinely has fewer ways to cause harm than it would with unrestricted access. But they have not implemented security. They have implemented a reduced-capability deployment — and if a prompt injection attack, a compromised orchestration layer, or a chained tool-call sequence exploits what remains, the architecture has no enforcement layer beneath the capability list to stop it.

The gap between these two postures is widening as agentic systems become more capable, more interconnected, and more deeply embedded in business-critical workflows. An agent restricted to “safe” capabilities in 2025 will encounter increasingly sophisticated adversarial inputs. A capability list is a static defense. An adversary is a dynamic problem. Structural security controls — identity enforcement, zero trust microsegmentation, dynamic authorization, policy-as-code, behavioral monitoring at the action layer — are the architectural response to a threat that evolves faster than any curated tool set can anticipate.

80%

of organizations report encountering risky behaviors from AI agents, including improper data exposure and unauthorized system access — in deployments that had already implemented capability restrictions. The restriction layer did not fail. The structural enforcement layer was absent. — SailPoint Technologies, AI Agents: The New Attack Surface, May 2025

Defining the Terms Precisely

The conflation of these two concepts is partly a vocabulary problem. Both are discussed under the umbrella of “AI agent security,” both involve limiting what agents do, and both are recommended by authoritative frameworks including OWASP LLM Top 10 and the NIST AI Risk Management Framework. The difference lies in where the limiting happens and what enforces it.

What It Is

Capability Restriction

Capability restriction is the practice of limiting the tools, permissions, functions, and access scope available to an AI agent at design or deployment time. It operates on the supply side: the agent is given less to work with, so there is less it can do wrong. Examples include read-only database tokens, curated MCP tool sets, restricted API key scopes, and the removal of high-risk integrations such as email, browser automation with live sessions, or write access to production systems.

Capability restriction is meaningful risk reduction. It is also static, design-time, and entirely dependent on the accuracy of the initial threat model. It does not enforce anything at runtime. If the agent is compromised, jailbroken, or manipulated into using a remaining capability in an unintended way, capability restriction has no response. There is no enforcement layer — only a smaller list of things that can go wrong.

Supply-Side Hygiene
What It Is

Structural Security Controls

Structural security controls are architectural, enforcement-layer mechanisms that govern how an agent interacts with systems regardless of what capabilities it nominally possesses. They operate beneath the prompt level — at the identity, network, protocol, and policy enforcement layers — and they enforce authorization decisions that the agent cannot reason, jailbreak, or prompt-inject its way around. The key question is not what the agent can do, but what the agent is authorized to do in this specific context, under this verified identity, at this moment.

Structural controls include non-human identity management with cryptographic verification, dynamic just-in-time authorization with no standing privileges, zero trust microsegmentation, policy-as-code with runtime enforcement, behavioral monitoring at the action layer, and scoped delegation tokens governed by RFC 8693 and OAuth 2.0 token exchange. These controls are enforced at the infrastructure layer. They do not depend on the agent’s compliance or the accuracy of its output.

Enforcement-Layer Architecture

Why the Substitution Fails: Five Structural Breakdowns

The core argument is not that capability restriction is useless — it is that it is being deployed as a primary security strategy for systems that require an enforcement architecture. The substitution fails in five distinct and compounding ways.

Why Restriction Fails Here

Prompt Injection Makes Restriction Incomplete

Capability restriction addresses what tools exist in the agent’s environment. It does not address whether an existing, authorized tool can be weaponized against its own intended purpose by injected instructions embedded in data the agent reads. An agent restricted to S3 read access can still be directed by a malicious document to exfiltrate contents to an attacker-controlled endpoint — if the agent has any external communication capability remaining. The “Lethal Trifecta” — sensitive data access, untrusted content exposure, and external communication ability — can be constructed from capabilities that each appear safe in isolation.

Simon Willison — Lethal Trifecta for AI Agents; martinfowler.com/articles/agentic-ai-security, 2025
What Structural Controls Provide

Protocol-Layer Enforcement Below the Prompt

Controls at the network, API protocol, and identity communication layer cannot be bypassed by social engineering of the LLM, because they do not involve natural language interpretation. A zero trust access decision evaluating whether a verified agent identity is authorized to reach a specific endpoint is not reading the agent’s output. It is enforcing a policy against a cryptographic identity. Even a successfully injected prompt that directs the agent toward unauthorized behavior will fail at the enforcement layer — because the agent lacks the standing privilege to execute the action, regardless of what it was instructed to do.

Xage Security — Why Zero Trust is Key to Securing AI, LLMs, Agentic AI, MCP Pipelines; xage.com, 2025
Why Restriction Fails Here

Static Restriction Meets Dynamic Execution

AI agents are invoked by users but execute actions independently, persist beyond the initiating interaction, and adapt behavior dynamically. Trust is evaluated at the moment the agent is deployed or a session is initiated. Execution — and risk — occurs later, across chained tool calls that were not individually anticipated at design time. An agent restricted to a safe-looking capability set at deployment time can accumulate effective privilege through sequential, individually-authorized actions that together produce an outcome the restriction was intended to prevent. Capability restriction cannot account for execution paths that do not exist at design time.

Zvelo — Agentic AI Security Exposes the Limits of Zero Trust; zvelo.com, February 2026
What Structural Controls Provide

Just-in-Time Authorization with No Standing Privilege

Dynamic authorization evaluates each access request in real time against the agent’s current verified identity, the context of the request, and the scope defined for the active task. Scoped delegation tokens issued via OAuth 2.0 token exchange (RFC 8693) grant the agent access it needs for a specific action window and expire. The agent has no standing privilege to exploit between tasks. An attacker who compromises the agent at step four of a twelve-step workflow gains only whatever JIT access was scoped for step four — not the accumulated permissions of a persistent service account.

CyberArk — Zero Trust for AI Agents: Delegation, Identity and Access Control; developer.cyberark.com, August 2025
Why Restriction Fails Here

Capability Restriction Is Not Least Privilege

These two concepts are frequently conflated, but they are architecturally distinct. Least privilege is an access control property: an identity is granted only the permissions required to perform a specific task, enforced at the authorization layer. Capability restriction is a deployment property: an agent is given fewer tools. A chatbot with a restricted tool set but super-user database credentials — capable of reading all customer accounts while filtering its output based on the requesting user — has undergone capability restriction. It has not been granted least privilege. The data is still reachable. The filtering is happening at the output layer, not the access control layer. Any technique that bypasses or manipulates the output filter exposes everything.

CMS Information Security — Zero Trust Identity Principles for AI-Based Services; security.cms.gov, 2025
What Structural Controls Provide

Identity-Aware Access Control at the Data Layer

Structural least privilege enforces access decisions at the resource layer — the database, the API endpoint, the file system — not at the agent’s output layer. The agent does not receive data it is not authorized to see; the authorization decision happens before the data is retrieved. Fine-grained, context-aware controls limit agent access based on real-time factors: the requesting user’s identity, the sensitivity classification of the data, the time window of the request, and the specific scope of the active task. This is not filtering. It is access control — and it holds regardless of what the agent’s output logic does or does not do with the result.

Xage Security — Zero Trust for AI Security: Identity-First Defense; xage.com, October 2025
Why Restriction Fails Here

Blast Radius Is Uncontrolled

Even in a deployment where capability restriction has been thoughtfully applied, a compromised or manipulated agent operating within its restricted capability set can still cause significant damage — because the restriction defines the ceiling of possible harm, not the floor. A well-restricted agent with read access to three data sources and one external API can still exfiltrate sensitive data from all three sources to the external API if it is successfully injected. The damage scales with the sensitivity of what remains accessible, not with the size of the capability list. Capability restriction reduces the ceiling. It does not limit blast radius once a capability is being misused.

CyberArk — Zero Trust for the Age of Autonomous AI Agents Part 2; bankinfosecurity.asia, 2025
What Structural Controls Provide

Architectural Containment and Microsegmentation

Zero trust microsegmentation treats every AI agent as its own isolated zone with explicit, cryptographically enforced rules about what it can communicate with. A compromised agent cannot reach systems outside its defined segment, regardless of what it has been instructed to do. Behavioral monitoring at the action layer — not the output layer — detects anomalous patterns before they compound: an agent making data retrieval calls at unusual frequency, attempting to reach endpoints outside its defined scope, or chaining actions in sequences that deviate from established behavioral baselines. Detection happens at the infrastructure layer, in real time, without requiring the agent’s output to be interpretable.

Cisco — Zero Trust in the Era of Agentic AI; blogs.cisco.com, September 2025

Capability restriction asks what we can take away. Structural security asks what we can enforce. Only one of those questions produces an architecture that holds when the agent is compromised, manipulated, or operating in ways its designers did not anticipate.

— Luminity Digital, Agentic AI Security Practice, March 2026

The Scale Argument: Why Restriction Cannot Grow with the Fleet

There is a practical dimension to this failure that becomes visible at enterprise scale. Capability restriction requires continuous human curation. Every new use case, every new MCP server, every new integration requires a security review that produces a curated capability list. That model functions at small scale. It collapses at the scale agentic AI actually reaches in production environments.

Machine identities now outnumber human identities by a ratio of approximately 45-to-1 in enterprise environments. Best practices for agentic system design recommend decomposing complex tasks into multiple smaller, specialized agents that operate in coordination — which means the number of distinct agent identities grows with the complexity of the workflows being automated. Each of those agents, in the capability-restriction model, requires its own curated tool set, maintained over time as use cases evolve and integrations change.

The Shadow AI Agent Problem

Capability restriction is only applied to agents that security teams know about. Enterprise deployments are already contending with what practitioners are calling Shadow AI Agents: agents deployed by business users through SaaS integrations, browser extensions, and productivity tools — often connecting to corporate data sources without IT visibility. An agent a security team has not inventoried cannot have its capabilities restricted. Structural security controls enforced at the identity and network layer apply to all agents by default: an agent without a registered, cryptographically verified identity is denied access. The control posture does not depend on advance knowledge of the agent’s existence.

Structural security controls scale through policy-as-code and automated identity governance. Every agent is required to register with a central machine identity provider before it can access any resource. Any agent that does not possess a cryptographically signed identity and a defined scope of work is denied network access by default. The security posture does not depend on a human having reviewed and restricted the agent’s capability set — it depends on the agent having a verified identity and an authorized scope, both of which are enforced at the infrastructure layer regardless of fleet size.

What Structural Security Controls Actually Look Like

The following is not a conceptual framework. It is a description of specific, implementable architectural components that currently exist, are deployed in production enterprise environments, and are documented in authoritative standards including NIST SP 800-207, the CSA Agentic Trust Framework, and the OWASP Top 10 for Agentic Applications 2026.

1. Non-Human Identity Management

Every agent must have a verified, auditable identity before it can access any resource. This goes beyond authentication to include the full identity chain: who owns this agent, what is its purpose, what capabilities does it claim to have, and is the underlying model code cryptographically bound to a trusted execution environment. Hardware-backed machine identity ensures that even if an agent’s logic is subverted, the identity itself remains bound to a verified state — making it possible to detect when an agent’s behavior deviates from the profile associated with its registered identity.

2. Dynamic Just-in-Time Authorization

Static role-based access control was designed for a world where roles changed infrequently. AI agents operate across multiple contexts within minutes. JIT authorization issues scoped credentials for each specific task window and revokes them on completion, leaving no standing privilege for an attacker to exploit between tasks. This is formalized in RFC 8693 token exchange — the agent receives a delegation token that represents the intersection of what the owning user is permitted to do and what the agent is authorized to do on their behalf, within a defined scope and time window.

3. Zero Trust Microsegmentation

Each agent is treated as its own isolated zone with explicit, cryptographically enforced rules about what endpoints it can reach and what systems it can communicate with. Traditional network segmentation creates large zones that agents routinely need to traverse. Microsegmentation enforces communication rules at the agent level. A compromised agent cannot reach systems outside its defined segment regardless of what instructions it receives, because the enforcement is at the network layer, not the application layer.

4. Policy-as-Code with Runtime Enforcement

Authorization rules are expressed as code — auditable, testable, version-controlled, and automatically enforced at runtime. Policy engines evaluate each agent action request against defined rules before the action is permitted. The MCP enterprise profile discussion underway in the community points toward publishing tool sensitivity metadata that policy engines can consume, enabling governance logic such as: only agents with a defined trust level, operating under a verified user consent scope, may invoke tools tagged with elevated sensitivity classifications. This is enforceable and auditable in ways that a capability list is not.

5. Behavioral Monitoring at the Action Layer

Production agentic security requires observability at the action layer — the APIs being called, the data being retrieved, the sequences of operations being executed — not only at the output layer. Behavioral baselines established during initial deployment define what normal agent operation looks like: expected data access patterns, typical tool call frequencies, normal sequences of action. Deviation from those baselines triggers detection and response mechanisms that do not depend on the agent’s output being interpretable. An agent that has been successfully injected with malicious instructions will change its behavior at the action layer before its output reveals the compromise.

What a Structurally Secure Agentic Deployment Requires in Practice

First, establish an agent registry before any production deployment. Every agent must be inventoried with a defined owner, purpose, data access requirements, and operational boundaries. This is the foundation for all downstream identity and authorization controls. Agents that cannot be inventoried cannot be governed.

Second, implement JIT authorization from the first deployment, not as a later hardening step. The pattern of granting standing service account credentials to agents and restricting their scope through capability lists is the pattern that creates the substitution problem. JIT authorization requires more engineering investment upfront and produces a fundamentally different security posture.

Third, build behavioral baselines at initial deployment and treat deviation as a detection signal. The question is not whether your agent is behaving correctly today — it is whether you will know when it is no longer behaving correctly in production. Baselines established at launch are the detection mechanism for injection attacks, model drift, and compromised orchestration layers.

Fourth, map compliance requirements to specific structural controls before architecture is finalized. GDPR Article 22, EU AI Act high-risk provisions, NIST AI RMF Govern and Measure functions, and internal governance frameworks all have direct implications for identity logging, audit trails, and authorization decision records. These requirements are far easier to satisfy when the structural controls are designed with them in mind than when retrofitted to a running fleet.

Capability restriction will remain a meaningful component of agentic AI security practice. The discipline of minimizing what an agent can access, providing only the tools required for a specific job, and avoiding the antipattern of granting broad permissions to hedge against future use cases — these are sound engineering practices and they reduce risk in measurable ways. The argument here is not against them. It is against the organizational pattern of performing capability restriction, labeling the result “secured,” and deploying into production without an enforcement architecture beneath it. That pattern mistakes a reduced attack surface for a defended one — and in an adversarial environment, the difference is the entire security posture.

Practitioner Takeaway

If your agentic deployment has a curated tool list but no cryptographic agent identity, no JIT authorization, no behavioral monitoring at the action layer, and no policy enforcement below the prompt level, you have not implemented agentic AI security. You have implemented a smaller attack surface. That is not nothing — but it is not a security architecture, and it will not hold against an adversary who understands the difference between what an agent is capable of and what an agent is structurally prevented from doing.

Capability Restriction vs. Structural Security Controls — March 2026

This post draws on OWASP Top 10 for Agentic Applications 2026, NIST SP 800-207 Zero Trust Architecture, RFC 8693 OAuth 2.0 Token Exchange, the CSA Agentic Trust Framework (February 2026), the AWS Agentic AI Security Scoping Matrix (November 2025), Simon Willison’s Lethal Trifecta framework, SailPoint AI Agents survey research, and current technical documentation from CyberArk, Cisco, Xage, Cerbos, and Zvelo. Full citations below.

References & Sources
Tags
Agentic AI Security Zero Trust Capability Restriction Structural Security Non-Human Identity JIT Authorization OWASP LLM Top 10 Prompt Injection Policy-as-Code MCP Security AI Governance NIST AI RMF RFC 8693 Microsegmentation

Share this: