MCP Is the USB-C of AI Tools. Most Teams Haven't Plugged In Yet.

Cover Image for MCP Is the USB-C of AI Tools. Most Teams Haven't Plugged In Yet.

Cover photo by Google DeepMind on Pexels

Three months ago, a team at a fintech built a custom integration to let their Claude-based assistant read from their Postgres database. Twelve engineering days. Last week, a different team did the same thing in an afternoon using Model Context Protocol. The fintech team's code is already technical debt.

MCP is 18 months old. It's already become the standard protocol for connecting AI models to external tools. Most engineering teams still treat it as an Anthropic-specific curiosity. That's a compounding mistake.

What MCP Actually Is (and Why the USB-C Analogy Holds)

Before Model Context Protocol, connecting a language model to an external tool required custom integration code for every tool-model pairing. Two models, five tools: ten integrations. Add a new model: five more integrations. Change an API: five updates. The combinatorial explosion was manageable when AI tooling was experimental. It isn't manageable in production.

MCP is a client-server protocol that standardizes the interface between models and tools. The model acts as a client. MCP servers expose tools, resources, and prompt templates through a standardized schema. Any model that speaks MCP can connect to any MCP server. One server, any compatible model.

Anthropic released the specification in November 2024 as open source. The adoption curve was steep: OpenAI announced support in March 2025. Google's Gemini team followed. Cursor ships MCP natively. VS Code Copilot, Windsurf, and Zed all added MCP support within the first quarter of 2025. The spec won before most teams knew there was a competition.

The USB-C comparison is imperfect but accurate at the right level of abstraction. Before USB-C, a laptop needed different cables for power, display, data transfer, and peripherals. USB-C didn't eliminate the underlying complexity — a USB-C port can still carry Thunderbolt or just USB 2.0 — but it standardized the physical and protocol interface. MCP does the same for LLM-to-tool connectivity. The underlying tools still have their own authentication, rate limits, and failure modes. MCP standardizes the interface your model speaks to them through.

What It Looks Like in Production

By April 2025, over 2,500 MCP servers had been indexed on GitHub. The ecosystem covers the obvious integrations — filesystem access, PostgreSQL, GitHub, Slack, email, web search, code execution — and expands into domain-specific tools daily. Anthropic's own catalog includes 40+ reference implementations. The infrastructure exists.

The production model is straightforward: you run MCP servers alongside your application. Your AI model discovers available tools through the server's schema, decides when to call them based on context, and receives structured results it can reason about. The model doesn't need to know that the database is Postgres, that the authentication is OAuth, or that the rate limit is 1,000 requests per minute. It calls a tool called query_database with parameters, gets back results.

Where this breaks down in practice mirrors traditional API integration: server stability, rate limits, authentication token refresh, and failure handling. MCP doesn't solve these — it standardizes the interface above them. A poorly configured MCP server fails in exactly the ways a poorly configured REST endpoint fails. The protocol is not magic.

The strategic implication, though, is real: teams building custom tool integrations today are building against a protocol that's actively being deprecated by the ecosystem. They're solving a problem that MCP already abstracts. Every custom integration is N engineers times M maintenance cycles — every model update, every tool API change, every new capability requires touching that code.

Where It Actually Gets Hard

MCP's Achilles heel isn't the protocol — it's tool schema design. A model can only use tools it can understand, and understanding comes from the schema description. A tool called get_data with a description of "retrieves data" is effectively invisible to a model that needs to decide whether to call it. Good MCP tool design requires the same discipline as good API design: precise naming, clear parameter descriptions, specific examples of what the tool returns.

This is where most teams underestimate the work. They stand up an MCP server quickly and wonder why the model doesn't use the tools correctly. The model isn't wrong — the schema is ambiguous. Fixing it requires iteration with the model itself: ask the model to explain why it chose not to call a specific tool, read the schema through that lens, revise.

Multi-server orchestration compounds the challenge. Production agentic systems often need 5 to 15 MCP servers — one for file access, one for code execution, one for database reads, one for web search, several for internal APIs. Managing discovery, authentication, and graceful failure across all of them is real engineering work MCP doesn't do for you. You still need circuit breakers, fallback behavior, and logging to understand why a tool call failed.

The "context" in Model Context Protocol deserves attention: MCP enables live data access for models that otherwise rely on static training cutoffs. A model connected to your Postgres database via MCP isn't guessing based on training data — it's reading your actual data. This is powerful and comes with the full scope of database access concerns: what data are you exposing, to which models, under what conditions, with what audit trail.

The Composability Argument

The teams sleeping on MCP aren't in danger of being unable to build agentic systems. They're in danger of building ones that are infinitely harder to maintain, swap, and extend.

The real value of MCP isn't that it makes individual integrations easier. It's that it makes the whole architecture composable. When you need to switch from Claude to GPT-4o for a specific task, you don't rewrite your tool layer — the MCP server already speaks both. When you add a new capability, you write one MCP server, not N integrations. When a tool changes its API, you update one server, not scattered integration code across your codebase.

This is the same argument that made REST APIs eventually win over custom protocols, and that made Docker containers eventually win over custom deployment scripts. Not "it's better in isolation" but "it's better when your system grows to 20 services and you need them to compose."

Multi-agent systems make this especially pointed. If you're building workflows where multiple agents hand off to each other, they need a shared tool layer. Without MCP, every agent has its own integration code. With MCP, every agent connects to the same servers.

The teams that understood REST early enough built things other teams couldn't. The same pattern is playing out now, faster.

What To Do With This

If your team isn't already using MCP: the barrier to entry is lower than it looks. Anthropic's reference implementations cover most common integrations. Setting up a filesystem or database MCP server is an afternoon of work, not a sprint.

The place to start is a single, high-value integration your AI feature needs right now. Build the MCP server for it. Spend time on the tool schema. Test it by asking the model to explain why it calls the tool in specific scenarios. Then look at whether the pattern generalizes to your next three integrations.

The harder question — which tools to expose, with what level of access, to which models — is an architecture question MCP doesn't answer for you. Treat MCP servers the way you treat service boundaries: define them carefully, don't expose more than necessary, audit access.

For teams building agentic AI into production products right now, MCP is not optional infrastructure. It's the abstraction layer the ecosystem is converging on. Every week you spend building below it is a week of work you'll eventually redo.

The teams that will compound the most value in AI product development aren't the ones with the best prompts. They're the ones that built on the right abstraction layer early enough to benefit from everything the ecosystem builds on top of it.


Related: The 35% Problem: Why Multi-Agent AI Workflows Collapse in Production