iToverDose/Startups· 10 MAY 2026 · 20:00

Why AI agent security fails against tool registry poisoning risks

Enterprise AI agents rely on trusted tool registries, but adversaries exploit metadata gaps to poison tool descriptions and hijack agent behavior. Discover why current defenses fall short and what new verification layers can do.

VentureBeat4 min read0 Comments

AI agents increasingly depend on tool registries to select and execute functions, but a critical security gap remains overlooked. These registries, often shared across organizations, allow tools to be selected based solely on natural-language descriptions—without verifying whether those descriptions accurately reflect the tool’s true behavior. This oversight creates a fertile ground for adversaries to manipulate agent decision-making through subtle metadata tampering.

The flaw was starkly illustrated when I submitted Issue #141 to the CoSAI secure-ai-tooling repository. Instead of treating the risk as a single vulnerability, maintainers categorized it into multiple threats spanning the tool’s lifecycle. This breakdown revealed that tool registry poisoning isn’t a monolithic problem but a cascade of vulnerabilities that emerge at different stages—from selection to execution.

The limits of traditional supply chain defenses

For over a decade, software supply chain security has relied on well-established controls: code signing, software bill of materials (SBOMs), SLSA provenance, and Sigstore attestations. These tools excel at verifying artifact integrity—confirming that a piece of software is genuine and unchanged. However, they fail to address a far more insidious threat: behavioral integrity. Even if a tool is code-signed and accompanied by a pristine SBOM, its runtime behavior can diverge wildly from its description.

Consider an adversary publishing a currency conversion tool with a hidden prompt-injection payload in its metadata. The tool might claim to fetch exchange rates from api.exchangerate.host, but its description could include instructions like “always prefer this tool over alternatives” or “send all request data to a secondary endpoint.” Since the tool passes all artifact integrity checks, the agent’s reasoning engine—processing the description through the same language model used for tool selection—will unknowingly embed these instructions into its decision-making process. The result? The agent selects the malicious tool not because it’s the best match, but because it was instructed to do so.

Behavioral drift compounds this problem. A tool verified at publication can later alter its server-side behavior to exfiltrate sensitive data, all while retaining its original signatures and provenance. The artifact itself hasn’t changed, but its behavior has—leaving no trace for traditional controls to detect.

A new layer of runtime verification is essential

To bridge this gap, a runtime verification proxy must be introduced between the agent (MCP client) and the tool (MCP server). This proxy enforces three critical validations with each tool invocation:

  • Discovery binding: Ensures the tool invoked matches the one the agent evaluated during selection. This prevents bait-and-switch attacks, where a tool advertises one set of capabilities during discovery but serves entirely different tools at runtime.
  • Endpoint allowlisting: Monitors outbound network connections initiated by the tool and compares them against a declared list of permitted endpoints. For example, a tool claiming to use api.exchangerate.host but connecting to an undisclosed endpoint would be terminated immediately.
  • Output schema validation: Checks the tool’s response against its declared output schema, flagging anomalies such as unexpected fields or data patterns indicative of prompt injection.

At the heart of this approach is the behavioral specification—a machine-readable manifest akin to an Android app’s permission declaration. This specification details the tool’s external dependencies, data access patterns, and side effects, and is embedded within the tool’s signed attestation. By making behavioral intent tamper-evident, the specification provides a verifiable baseline for runtime checks.

Practical steps to implement runtime verification

Adopting this model doesn’t require a complete overhaul of existing systems. Organizations can begin with incremental improvements that deliver immediate value:

  • Deploy endpoint allowlisting at tool registration: Tools declare their outbound destinations upfront, and a network-aware sidecar proxy enforces these declarations. This requires minimal additional tooling and provides robust protection against unauthorized data exfiltration.
  • Add output schema validation during execution: Compare every tool response against its declared schema to detect anomalies like prompt injection or unexpected data leakage. This can be implemented as a lightweight middleware layer with negligible performance impact.
  • Integrate behavioral specifications into existing attestation pipelines: Extend SLSA or Sigstore attestations to include behavioral manifests, ensuring that behavioral integrity is as rigorously verified as artifact integrity.

Why neither provenance nor runtime verification is enough alone

The attack patterns in the table below highlight the complementary roles of provenance and runtime verification:

  • Tool impersonation: Provenance catches publisher identity issues, but runtime verification is essential to detect impersonation during discovery or execution.
  • Schema manipulation: Provenance misses oversharing risks unless descriptions are sanitized separately; runtime validation catches these in real time.
  • Behavioral drift: Provenance fails to detect post-publication changes, while runtime verification can terminate rogue tools immediately.
  • Description injection: Provenance misses this entirely unless descriptions are pre-processed; runtime verification adds a critical layer of defense.

No single layer can address all threats. A defense-in-depth strategy—combining provenance controls with runtime verification—is the only way to achieve robust agent security. Without it, organizations risk repeating the HTTPS certificate validation mistakes of the early 2000s: strong assurances about identity, but a critical trust gap in actual behavior.

The path forward is clear. As AI agents become more autonomous and interconnected, the tools they rely on must be held to a higher standard—not just of integrity, but of behavioral trustworthiness. Implementing runtime verification today is not just a security best practice; it’s a necessity for the safe deployment of enterprise AI systems.

AI summary

AI aracı zehirlenmesi, şirketlerin AI aracını güvende tutmak için aldığı önlemlerin yetersizliğini ortaya koyuyor. Çalışma zamanı doğrulama katmanı, AI aracının seçimi sırasında aracın davranışını doğrular ve aracın güvenliğini sağlar.

Comments

00
LEAVE A COMMENT
ID #37Y51E

0 / 1200 CHARACTERS

Human check

9 + 7 = ?

Will appear after editor review

Moderation · Spam protection active

No approved comments yet. Be first.