Skip to main content

Deprecation Notices Don't Work for AI Agents

40% of enterprise apps will include AI agents by 2026. MCP downloads hit 8 million in five months. When autonomous agents become your primary API consumers, the traditional versioning contract — email notice, six-month window, hard sunset — breaks down completely.

  • api-versioning
  • ai-agents
  • mcp
  • gateway
  • operations
Zerq team

Gartner projected that 40% of enterprise software applications would include task-specific AI agents by 2026, up from under 5% in 2024. MCP — the protocol that connects those agents to APIs — grew from 100,000 downloads in November 2024 to 8 million by April 2025. That is 8,000% growth in five months.

The tooling for AI agent connectivity moved fast. The operational practices for managing it did not.

One gap that has gone mostly unexamined: API versioning. The model that worked for human developers — send a deprecation email, give six months, watch traffic migrate, sunset — was already imperfect. For AI agents, it does not work at all.


The old contract assumed a human on the other side

The traditional versioning lifecycle is a social contract. You announce a deprecation. A developer reads the email, creates a ticket, ships a migration. Traffic on v1 declines. You sunset.

Every step assumes a human is paying attention. An AI agent calling your API has a credential. It does not have an inbox.

When you send a migration notice to the developer who originally provisioned the API key, there is no guarantee that developer is still responsible for the agent, still at the company, or has any visibility into what the agent is actually doing at runtime. The agent is autonomous. It calls what it was configured to call.

This is not hypothetical. Nordic APIs identified it directly in early 2025: "When an MCP tool that wraps an API is prone to becoming obsolete in a matter of months — or weeks — the connective tissue that MCP offers can quickly become a problem." Their analysis, titled "The Weak Point in MCP Nobody's Talking About: API Versioning", called out exactly why the current approach fails in an agent-first world.


Four ways it breaks

1. MCP tools hardcode paths. A common pattern is wrapping an API in an MCP tool definition that specifies the exact route, method, and parameter shape. That definition lives in a repository and gets deployed into an agent's tool list. Updating it requires a code change, a deployment, and someone noticing the deprecation in the first place. Separately, a Medium analysis found the compounding effect: "Your MCP Server's Tool Descriptions Changed Last Night. Nobody Noticed." — tool description changes silently alter agent decision-making with no audit trail.

2. Agent workflows cannot be fully regression-tested. A human developer can run their integration against staging and verify the migration works. An agent's tool usage is emergent — it decides at runtime which tools to call and with what parameters. You cannot fully test an agent's migration path the way you can test a conventional service. The only way to know an agent still works after a version change is to run it.

3. Deprecated endpoints become security liabilities. The Stripe incident in 2025 demonstrated this clearly: attackers found a legacy endpoint (/v1/sources) that was still connected to backend payment systems but lacked the security controls of modern APIs — advanced rate limiting, fraud detection, current authentication requirements. The deprecated route had never been hardened because it was "being retired anyway." It affected 49 online retailers. An endpoint that is deprecated but still reachable is not a neutral endpoint. It is an exposed one.

4. You have no visibility into who is still on v1. Most teams cannot answer "which clients are still calling the old version?" with a quick query. They find out when the sunset breaks something. The Cloud Security Alliance's 2025 analysis found that 51% of developers cite unauthorized or excessive API calls from AI agents as their top security concern. Untracked agent calls to deprecated routes feed directly into that risk.


What actually works in 2026

The answer is not better communication. It is moving versioning from a social contract to an operational one.

Per-client version pinning

Rather than a single sunset date for all clients, assign each client a version policy at the gateway. A legacy agent calling /v1/payments gets pinned to v1 routing. Every new integration uses v2. The migration timeline becomes per-client, not per-API — which means well-maintained integrations can move fast while others get the time they actually need.

Gateway-level request rewriting

For breaking changes that are structural but not semantic — a renamed field, a changed path segment, a different HTTP method — the gateway can translate. The agent sends a v1-shaped request; the gateway rewrites it before it reaches the upstream. This is not a permanent solution, but it buys time to update agent tooling without causing an incident. The upstream can drop v1 entirely while lagging agents continue to be served.

Traffic audit by version

Every request should log the API version, client ID, actor type (human, agent, copilot), and endpoint. With that data, a sunset decision becomes a query:

db.audit_logs.aggregate([
  { $match: { api_version: "v1", timestamp: { $gte: thirtyDaysAgo } } },
  { $group: { _id: "$client_id", call_count: { $sum: 1 } } },
  { $sort: { call_count: -1 } }
])

Run this before you sunset anything. Empty result: sunset with confidence. Six clients still calling v1 — three of which are agents you have never heard of — means you have work to do first, and you found out before the incident rather than during it.

The Theneo 2026 API management guide found that organisations implementing proactive change management report 70% fewer update-related incidents. The difference between proactive and reactive here is having version-tagged audit data before a decision, not after.

Zero-traffic sunset, not calendar sunset

The calendar-based sunset was invented because teams had no better signal. With real-time traffic data broken out by client and version, you do not need an arbitrary deadline. You sunset when traffic reaches zero.

That might be faster than six months for active integrations. It might be longer for agent-driven ones. The data tells you when it is safe.


The paradigm shift is already underway

The AI Infrastructure Link's 2026 versioning analysis put it plainly: "In 2026, API versioning has transitioned from a technical best practice into a core business discipline... Enterprise AI agents depend on stable APIs for real-time data, and inconsistent versioning can cause these systems to generate erroneous outputs, eroding trust in AI-driven decision-making."

The teams navigating this well share a common approach: they treat versioning as a gateway concern — enforced by policy, audited by data, per-client by default. They do not rely on a notice in a changelog that an autonomous agent will never read.


Zerq's per-partner routing configuration supports version pinning at the client level. Every request is logged to your MongoDB instance with version, actor type, and client identity — so when you're ready to retire an old version, the audit log tells you whether it's safe to do so.

See how the gateway handles routing →