You installed 5 MCP servers. Your context window just lost 40,000 tokens. And you had no idea.
The Hidden Cost Nobody Talks About
MCP (Model Context Protocol) is transforming how AI assistants interact with the world. But there's a cost nobody's measuring: every MCP tool silently eats your context window.
Here's the math that should worry you:
- Each MCP tool schema costs 550–1,400 tokens just to exist
- A server with 50 tools burns 30,000+ tokens before you type a single word
- That's context window space you're paying for — in money, speed, and quality
Real Numbers From Real Developers
This isn't theoretical. Here's what the community has discovered:
GitHub MCP Server: One user reported their token usage jumped from 34K to 80K tokens — a 135% increase — just by adding the GitHub MCP server. (source)
The 67K Token Tax: A developer measured 67,000 tokens consumed by just 4 MCP servers before any conversation even started. That's roughly 50 pages of text worth of context, gone. (source)
Context7 (240K weekly downloads): One of the most popular MCP servers has tool descriptions costing 3x more tokens than necessary. Bloated schemas are everywhere. (source)
Why This Matters More Than You Think
Context window space isn't free. When MCP tools eat your context:
- Your AI gets dumber — less room for your actual conversation means worse reasoning
- You hit rate limits faster — more tokens per request = fewer requests per minute
- You pay more — API users pay per token, and those schema tokens add up fast
- Responses slow down — more input tokens = longer processing time
The worst part? Most developers have zero visibility into this. You can't optimize what you can't measure.
Introducing mcp-checkup
I built mcp-checkup to solve exactly this problem. One command:
npx mcp-checkup
That's it. It installs as an MCP server and gives your AI assistant these capabilities:
What It Does
analyze_servers — Scans your MCP config and measures the token cost of every installed server. Auto-detects Claude Desktop, Claude Code, and Cursor configs.
analyze_tools — Deep-dives into a specific server: per-tool token costs, grades, bloated descriptions, optimization suggestions.
find_duplicates — Finds duplicate tool names across servers. Redundant tools waste context for zero benefit.
generate_report — Full markdown health report with grades, costs, duplicates, and actionable recommendations.
The Grading System
Every tool and server gets a letter grade:
| Grade | Per-Tool Tokens | Per-Server Tokens | Meaning |
|---|---|---|---|
| A | ≤100 | ≤500 | Lean and efficient |
| B | ≤300 | ≤1,500 | Reasonable |
| C | ≤600 | ≤3,000 | Could be optimized |
| D | ≤1,000 | ≤6,000 | Bloated |
| F | >1,000 | >6,000 | Major context drain |
How to Use It
1. Quick setup
Add to your Claude Desktop or Claude Code config:
{
"mcpServers": {
"mcp-checkup": {
"command": "npx",
"args": ["-y", "mcp-checkup"]
}
}
}
Or for Cursor, add to .cursor/mcp.json.
2. Ask your AI assistant
"Run an MCP health check on my setup"
"Which of my MCP servers costs the most tokens?"
"Generate a full MCP health report"
3. Get actionable results
You'll see a report showing each server's grade, token cost, tool count, duplicates, and optimization recommendations. Now you know exactly where the bloat is and can make informed decisions about which servers to keep, remove, or replace.
The Bigger Picture
As the MCP ecosystem explodes (there are already thousands of servers), context window management is becoming a real engineering problem. The developers building MCP servers need to think about schema efficiency, and the developers using them need tools to measure the cost.
Some practical tips beyond mcp-checkup:
- Audit regularly — your token budget changes every time you add/remove a server
- Prefer focused servers — a server with 5 well-designed tools beats one with 50 bloated ones
- Check for duplicates — multiple servers often register similar tools
- Read the schemas — some servers include entire documentation in tool descriptions
Links
- npm: npmjs.com/package/mcp-checkup
- GitHub: github.com/yifanyifan897645/mcp-checkup
- License: MIT
If you've been wondering why your AI assistant seems to perform worse with more MCP servers installed — now you know why. Measure first, optimize second.
Drop a comment if you run mcp-checkup and find any surprising results. I'd love to hear what the community discovers.
Top comments (0)