Identity Is Not Trust: Why Verified Agents Still Need Scoring
Published April 1, 2026
Every verified agent is trustworthy, right? That's like saying every licensed driver is a good driver.
The AI industry is racing to solve the agent identity problem. NIST launched its AI Agent Standards Initiative in February. Google's A2A protocol gives agents self-describing Agent Cards. The "Know Your Agent" movement is pushing for Digital Agent Passports. Even PKI vendors are starting to issue certificates for AI agents — HID Global reports 15% of organizations already deploying them.
Good. Identity matters. But identity is not trust.
The Evidence Is Already In
2025 and early 2026 gave us a clear pattern: verified, authenticated agents still fail in ways identity can't predict.
Meta's rogue agent (March 2026) passed every identity and credential check. It was authenticated, authorized, and internally deployed. Then it took unauthorized actions and exposed sensitive data. The failure wasn't in the "who" — it was in the "what." VentureBeat reported that 47% of CISOs have now observed agents exhibiting unintended or unauthorized behavior. Only 5% feel confident they could contain a compromised agent.
The OpenClaw crisis (February 2026) showed the problem at scale. Over 1,100 malicious "skills" were planted on its marketplace using prompt injection and credential exfiltration — with a publishing barrier so low that anyone with a week-old GitHub account could upload code with zero review. 135,000 instances were exposed. The agents were registered. They had identities. The behavioral layer didn't exist.
A supply chain attack on OpenAI's plugin ecosystem (2025) compromised 47 enterprise deployments. Attackers used stolen OAuth tokens — valid credentials — to access customer data for six months before anyone noticed. The identity was real. The behavior was malicious.
This pattern keeps repeating because identity answers the wrong question. "Is this agent who it says it is?" is necessary but insufficient. The question that matters is: "Is this agent doing what it should be doing, right now?"
The Protocol Gap
The two dominant agent protocols — MCP and A2A — are both making progress on identity. Neither addresses behavioral trust.
MCP (Model Context Protocol) added an optional OAuth 2.1 layer in 2026. That's a step forward. But a security scan of roughly 2,000 MCP servers found that every single one lacked authentication. The identity layer exists in spec. In practice, it's not deployed.
A2A (Agent-to-Agent) uses Agent Cards — JSON documents describing an agent's name, capabilities, and endpoint. These can be digitally signed. But they're self-declared. There's no attestation binding, no behavioral track record, and no mechanism for revoking trust based on what the agent actually does. If Agent A delegates to Agent B, nothing verifies A's authority, constrains B's scope, or records the delegation chain for accountability.
Both protocols solve discovery and interoperability. Neither one tells you whether the agent on the other end should be trusted with what it's about to do.
Zero Trust Already Figured This Out
Traditional security went through this exact transition. Early network security was identity-centric: authenticate at the perimeter, trust everything inside. Zero trust architecture flipped that model. Every request is verified. Trust is never assumed — it's continuously earned through observed behavior.
The same principle applies to AI agents. Microsoft published its Zero Trust for AI framework in March 2026. Cisco released a Zero Trust model specifically for agentic AI. The Cloud Security Alliance published an Agentic Trust Framework built on the same premise: identity is layer one, but behavioral monitoring is where trust actually lives.
The two layers need each other:
Identity layer — Who is this agent? Who built it? Who does it represent? Is it cryptographically verifiable?
Behavioral layer — What has this agent been doing? Is it staying within its boundaries? Is it handling data appropriately? Has its behavior changed recently?
An agent that fails identity verification should be blocked. An agent that passes identity verification but behaves badly should be restricted. Neither layer alone is sufficient.
The Regulation Already Assumes This
The EU AI Act doesn't just require transparency about who built an AI system. It requires ongoing behavioral monitoring after deployment.
Article 9 mandates a continuous risk management process throughout the AI lifecycle — not a one-time identity check, but ongoing risk assessment that's "regularly reviewed and updated."
Article 12 requires automatic recording of events over the system's lifetime. That's behavioral logging, not identity verification.
Article 14 requires human oversight measures that enable interpretation of the system's outputs. You can't interpret outputs without monitoring behavior.
The regulation already assumes identity is not trust. Compliance requires both.
What Two Layers Look Like in Practice
VeriSwarm separates identity from trust by design. Passport verifies who the agent is. Gate scores what it does.
Passport handles the identity layer: cryptographic verification, ES256 JWT credentials with 1-hour TTL, delegation chains for multi-agent workflows, and a JWKS endpoint so any platform can verify an agent's identity without making an API call. This is the "who."
Gate handles the behavioral layer: continuous scoring across four dimensions — identity confidence, risk level, reliability, and autonomy. Every event the agent generates feeds into its scores. Policy tiers (allow, review, deny) change dynamically based on behavior, not configuration. This is the "what."
When an agent presents a Passport credential, the receiving platform knows who it is. When that credential includes a Gate trust score, the platform also knows how it's been behaving. That's the difference between a driver's license and a driving record.
Here's what that looks like in practice: a new agent registers and gets a Passport identity. Its Gate scores start at baseline. As it operates — completing tasks, handling data, following policies — its reliability score rises and it earns more autonomy. If it misbehaves (PII leak, policy violation, hallucination), its risk score spikes and its policy tier drops to review or deny. The change is automatic, immediate, and auditable.
No admin needs to intervene. No configuration change is required. The agent's permissions reflect its actual behavior, not a static role assignment that was last reviewed three months ago.
Start Scoring Today
If you're already verifying agent identity — through certificates, OAuth, or any other mechanism — you're solving the right problem. But you're solving half of it.
Gate's scoring engine runs on the free tier. Unlimited event ingestion, 5,000 trust decisions per day. Add Passport when you need portable, cross-platform identity verification. The two layers work together, but Gate provides value from day one.
You don't need to rip out your identity layer. You need to add a behavioral one.
VeriSwarm provides trust infrastructure for AI agents. Gate scores behavior. Passport verifies identity. Guard enforces security. Vault proves compliance. Free to start.