<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Riddhesh</title>
    <description>The latest articles on DEV Community by Riddhesh (@riddhesh).</description>
    <link>https://dev.to/riddhesh</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1251181%2F08edd1cf-34f0-4e1c-b2df-130a19bff5e1.webp</url>
      <title>DEV Community: Riddhesh</title>
      <link>https://dev.to/riddhesh</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/riddhesh"/>
    <language>en</language>
    <item>
      <title>Should You Be Building on MCP in 2026?</title>
      <dc:creator>Riddhesh</dc:creator>
      <pubDate>Wed, 29 Apr 2026 10:06:06 +0000</pubDate>
      <link>https://dev.to/riddhesh/should-you-be-building-on-mcp-in-2026-47lc</link>
      <guid>https://dev.to/riddhesh/should-you-be-building-on-mcp-in-2026-47lc</guid>
      <description>&lt;h2&gt;
  
  
  What Is MCP (Model Context Protocol) and Why Does It Matter?
&lt;/h2&gt;

&lt;p&gt;Before 2024, connecting an AI agent to external tools meant writing custom integration code for every single combination. GitHub connector for Claude, a different one for GPT-4, another for your internal LLM. Each model, each tool, and each team has its own bespoke glue layer.&lt;/p&gt;

&lt;p&gt;That is the N×M problem. N AI models multiplied by M tools equals N×M custom integrations to build, maintain, and debug. For a team running three models across ten internal tools, that is thirty separate connectors. Every new model you add multiplies the work.&lt;/p&gt;

&lt;p&gt;Model context protocol collapses that to N+M. Each AI model implements the MCP client protocol once. Each tool or data source exposes an MCP server once. Any client can talk to any server without additional integration code.&lt;/p&gt;

&lt;p&gt;The USB-C analogy that circulates in every MCP explainer is accurate for once. One protocol, one port, everything connects. But here is the part most explainers leave out: MCP does not replace REST APIs. It sits above them. Your existing APIs still do the actual work. &lt;/p&gt;

&lt;p&gt;MCP is the orchestration layer that makes those APIs intelligible to an AI agent at runtime, not at build time.&lt;br&gt;
That distinction matters when you are making architecture decisions.&lt;/p&gt;

&lt;h2&gt;
  
  
  How MCP Took Over in 18 Months: The Adoption Timeline
&lt;/h2&gt;

&lt;p&gt;The adoption curve for MCP is unlike most open standards. It did not take years to reach critical mass. It took one.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Milestone&lt;/th&gt;
&lt;th&gt;Date&lt;/th&gt;
&lt;th&gt;Significance&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Anthropic launches MCP&lt;/td&gt;
&lt;td&gt;November 2024&lt;/td&gt;
&lt;td&gt;Open standard released with TypeScript and Python SDKs&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;OpenAI formally adopts&lt;/td&gt;
&lt;td&gt;March 2025&lt;/td&gt;
&lt;td&gt;Cross-vendor standardization confirmed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Microsoft Copilot Studio integration&lt;/td&gt;
&lt;td&gt;July 2025&lt;/td&gt;
&lt;td&gt;Enterprise channel opens&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;AWS adds support&lt;/td&gt;
&lt;td&gt;November 2025&lt;/td&gt;
&lt;td&gt;Cloud infrastructure layer adoption complete&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Donated to Linux Foundation (AAIF)&lt;/td&gt;
&lt;td&gt;December 2025&lt;/td&gt;
&lt;td&gt;Single-vendor dependency permanently removed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;97M monthly SDK downloads, 10K+ active public servers&lt;/td&gt;
&lt;td&gt;March 2026&lt;/td&gt;
&lt;td&gt;De facto infrastructure status&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;That last milestone is the one that changes the build decision. This is not a framework winning a competition. It is a standard that has already won. When OpenAI, Google, Microsoft, and AWS all implement the same protocol and donate governance to a neutral foundation, the question of whether MCP becomes the standard is closed. It already is.&lt;/p&gt;

&lt;p&gt;According to &lt;a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025" rel="noopener noreferrer"&gt;enterprise AI adoption data&lt;/a&gt;, 40% of enterprise applications will embed task-specific AI agents by the end of 2026, up from less than 5% in 2025. Every one of those agents needs to talk to tools. MCP is how they do it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What MCP Actually Solves for Developers
&lt;/h2&gt;

&lt;h3&gt;
  
  
  The N×M Integration Problem
&lt;/h3&gt;

&lt;p&gt;Industry reports indicate that connecting a legacy service via traditional REST APIs consumed three to five days of senior developer time per integration, before factoring in ongoing maintenance.&lt;/p&gt;

&lt;p&gt;Adopting MCP can reduce initial integration development time by up to 30% and lower ongoing maintenance costs by up to 25%, simply by eliminating the need to write custom connector code for every new AI platform.&lt;/p&gt;

&lt;p&gt;We have felt this directly. Before MCP in our stack, adding a new data source to an agent meant a full day of wiring: auth handling, response normalization, error mapping, and testing the edge cases. With MCP, a server that already exists connects in minutes.&lt;/p&gt;

&lt;h3&gt;
  
  
  Dynamic Tool Discovery
&lt;/h3&gt;

&lt;p&gt;This is the architectural shift most developers underestimate until they have built a real agentic system.&lt;/p&gt;

&lt;p&gt;Traditional REST assumes the client knows exactly which endpoint to call. A developer reads the docs, hardcodes the request, and ships it. That works fine for software calling software.&lt;/p&gt;

&lt;p&gt;AI agents operate differently. They need to ask at runtime, "What can I do here?" MCP servers answer that question through a tools/list call. The agent discovers available capabilities, reasons about which to use, and chains them across multiple steps all without any of that logic being hardcoded at build time.&lt;/p&gt;

&lt;h3&gt;
  
  
  Stateful Sessions vs Stateless APIs
&lt;/h3&gt;

&lt;p&gt;REST is stateless by design. Every request carries its full context. For human-written software, this is a feature it makes APIs predictable and easy to cache.&lt;/p&gt;

&lt;p&gt;For multi-step agent tasks, being stateless is a liability. When an agent is working through a sequence list of open PRs, summarize each, flag the stale ones, and close them if the context from step one is still available in step four.&lt;/p&gt;

&lt;p&gt;MCP sessions maintain that state. No re-authentication, no re-sending context, no re-establishing what the agent already knows.&lt;/p&gt;

&lt;h2&gt;
  
  
  MCP vs REST API vs Function Calling: Which One Do You Actually Need?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpp93naxp1u65bkmnkvig.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpp93naxp1u65bkmnkvig.png" alt="Hidden costs of MCP in production diagram showing security risks, token overhead, debugging complexity, and infrastructure impact on budget" width="800" height="533"&gt;&lt;/a&gt;&lt;br&gt;
This is the comparison most teams need before making an architecture decision, and it is the one nobody writes cleanly.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Approach&lt;/th&gt;
&lt;th&gt;Designed For&lt;/th&gt;
&lt;th&gt;State&lt;/th&gt;
&lt;th&gt;Tool Discovery&lt;/th&gt;
&lt;th&gt;Best Use Case&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;REST API&lt;/td&gt;
&lt;td&gt;Developer-written software&lt;/td&gt;
&lt;td&gt;Stateless&lt;/td&gt;
&lt;td&gt;Manual, hardcoded&lt;/td&gt;
&lt;td&gt;Direct integrations, high-volume deterministic operations&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Function Calling (native)&lt;/td&gt;
&lt;td&gt;Single-model tool use&lt;/td&gt;
&lt;td&gt;Per-call context&lt;/td&gt;
&lt;td&gt;Defined in system prompt&lt;/td&gt;
&lt;td&gt;Simple tool use within one model, known toolset&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;MCP&lt;/td&gt;
&lt;td&gt;AI agents, multi-model systems&lt;/td&gt;
&lt;td&gt;Stateful sessions&lt;/td&gt;
&lt;td&gt;Dynamic, runtime&lt;/td&gt;
&lt;td&gt;Multi-tool agentic workflows, cross-platform agent systems&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h3&gt;
  
  
  When REST still wins
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://code-b.dev/blog/setting-up-payment-apis" rel="noopener noreferrer"&gt;Payment processing&lt;/a&gt;, high-frequency deterministic operations, any integration where a human writes the calling code. REST is also significantly more mature on security OAuth 2.0, JWT, mTLS, and API gateway patterns have been battle-tested for over a decade.&lt;/p&gt;

&lt;h3&gt;
  
  
  When function calling is enough
&lt;/h3&gt;

&lt;p&gt;If you are using one model and your toolset is small and stable, native function calling is simpler. There is no server to run, no session to manage, and debugging is straightforward. Do not introduce MCP complexity for a use case that two tool definitions in a system prompt can handle.&lt;/p&gt;

&lt;h3&gt;
  
  
  When MCP earns its complexity
&lt;/h3&gt;

&lt;p&gt;More than three to five integrations, a need for dynamic tool discovery, cross-platform agent deployments where the same tools need to work with multiple AI models, or enterprise environments where centralized governance of agent actions is a requirement.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Problems With MCP in Production Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mj31zk1tld12z3d4d7h.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mj31zk1tld12z3d4d7h.png" alt="MCP vs REST API vs Function Calling comparison diagram showing differences in state management, tool discovery, and use cases for AI systems" width="800" height="533"&gt;&lt;/a&gt;&lt;br&gt;
MCP works great in demos. The production reality is messier, and most content about MCP right now skips the parts that will actually cost you time and money.&lt;/p&gt;

&lt;h3&gt;
  
  
  Security Is Genuinely Immature
&lt;/h3&gt;

&lt;p&gt;This is the section most MCP posts omit. It should not be omitted.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.pomerium.com/blog/mcp-server-security-risks-what-development-teams-need-to-know-in-2026" rel="noopener noreferrer"&gt;Security research&lt;/a&gt; covering 2,614 MCP implementations found that 82% had file operation vulnerabilities to path traversal attacks. Two-thirds had some form of code injection risk. Over a third were susceptible to command injection. These are not theoretical every category has confirmed CVEs with public exploits.&lt;/p&gt;

&lt;p&gt;The first two months of 2026 alone saw 30+ MCP-specific CVE filings from researchers at Check Point, Invariant Labs, and Adversa AI.&lt;br&gt;
The root cause is architectural. &lt;/p&gt;

&lt;p&gt;The MCP specification does not include built-in authentication or authorization. Every server you deploy inherits whatever permissions it has been granted. Every agent request flows through without verification unless you add controls externally, and most teams do not add them correctly or at all.&lt;/p&gt;

&lt;p&gt;Tool poisoning is a specific risk worth calling out. Malicious tool descriptions embedded in an MCP server's manifest can inject hidden instructions that the LLM reads and obeys without the user ever seeing them. Unlike prompt injection through user input, this attack is embedded in the protocol layer itself.&lt;/p&gt;

&lt;p&gt;Supply chain risk is real and has already caused incidents. A malicious package impersonating a legitimate email service was uploaded to an MCP registry, quietly exfiltrating API keys from developers who installed it.&lt;/p&gt;

&lt;p&gt;According to Gartner's 2026 security predictions, 25% of enterprise GenAI applications will experience at least five minor security incidents per year by 2028. MCP's current security posture, no built-in auth, and inconsistent registry vetting are evolving species and a direct contributing factor if teams treat it like a mature standard when it is not.&lt;/p&gt;

&lt;h3&gt;
  
  
  Context Window Bloat
&lt;/h3&gt;

&lt;p&gt;Connecting ten MCP servers with five tools each burns thousands of tokens before the agent does anything useful. Every tool schema loads into the context window at session initialization. On a 128K context window, that overhead is a real tax on both cost and latency that compounds with every additional server you connect.&lt;/p&gt;

&lt;h3&gt;
  
  
  Debugging Is Harder Than With REST
&lt;/h3&gt;

&lt;p&gt;When something goes wrong with a REST API call, you reproduce it with a curl command. Copy the request, run it, and inspect the response. Deterministic and fast.&lt;/p&gt;

&lt;p&gt;With MCP, a failure involves reading JSON-RPC transport logs, verifying the server process is still running, checking whether session state was corrupted, and determining whether the tool schema was cached incorrectly. &lt;/p&gt;

&lt;p&gt;Stateful sessions mean failures are harder to reproduce in isolation. There is no mature equivalent of Postman for MCP debugging yet.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Ecosystem Is Still Maturing
&lt;/h3&gt;

&lt;p&gt;Many public MCP servers still accept unauthenticated calls. Auth implementation quality varies dramatically. Registry vetting is inconsistent. The developer community is aware and moving to address it, but that is not the same as solved. &lt;/p&gt;

&lt;p&gt;If you are connecting third-party MCP servers to production systems today, you are accepting risk that has no industry-standard mitigation yet.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Anthropic's MCP Strategy Actually Signals
&lt;/h2&gt;

&lt;p&gt;The decision to donate MCP to the Linux Foundation under the Agentic AI Infrastructure Foundation, co-founded with Block and OpenAI with participation from Google, Microsoft, AWS, and Cloudflare, was the move that matters most.&lt;/p&gt;

&lt;p&gt;This was not a marketing decision. Donating governance to a neutral foundation removes the single-vendor risk that would otherwise limit enterprise adoption. &lt;/p&gt;

&lt;p&gt;It is the same structural move that made Kubernetes safe to bet on. MCP now sits alongside Kubernetes and PyTorch in the Linux Foundation portfolio a signal that carries real weight in enterprise architecture decisions.&lt;/p&gt;

&lt;p&gt;Anthropic's Claude Code integrates MCP natively, with explicit approval required for tool calls. That human-in-the-loop default is the right security posture. It reflects what a mature MCP implementation should look like not the "connect everything and let the agent run" approach that characterizes most early-stage deployments.&lt;/p&gt;

&lt;p&gt;MCP becomes invisible plumbing. The same way you do not think about TCP when you make a network request, the goal is that you will not think about MCP when your agent calls a tool. That abstraction layer is years away from being seamless. But the architectural bet is sound.&lt;/p&gt;

&lt;h2&gt;
  
  
  When You SHOULD Build on MCP in 2026
&lt;/h2&gt;

&lt;p&gt;Build on MCP when these conditions are true:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You are building multi-tool agentic workflows where the agent needs to discover and chain tools dynamically rather than execute a fixed sequence&lt;/li&gt;
&lt;li&gt;Your team manages more than three to five external integrations for AI and the custom connector maintenance cost is already visible&lt;/li&gt;
&lt;li&gt;You are building developer tooling that needs to plug into Claude Code, Cursor, or other MCP-native AI environments your users will expect it&lt;/li&gt;
&lt;li&gt;You need centralized audit trails and governance across agent actions, which MCP's session model enables more cleanly than ad-hoc API wiring&lt;/li&gt;
&lt;li&gt;You are building for enterprise deployments where the same agent tools need to work across multiple AI platforms without rebuilding connectors for each&lt;/li&gt;
&lt;li&gt;You want to stop rebuilding integrations every time a new model provider is adopted by your customers&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  When You SHOULD NOT Build on MCP
&lt;/h2&gt;

&lt;p&gt;Stop before adding MCP complexity when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You have one or two integrations and a direct API call is cleaner do not over-architect a simple problem&lt;/li&gt;
&lt;li&gt;Your use case is deterministic and does not require dynamic tool discovery at runtime&lt;/li&gt;
&lt;li&gt;Your security posture requires proven, audited auth patterns REST with OAuth 2.0 is significantly more mature right now&lt;/li&gt;
&lt;li&gt;You are in early prototyping and the overhead of running MCP servers and managing sessions slows down your ability to validate the idea first&lt;/li&gt;
&lt;li&gt;Your team has no prior experience with JSON-RPC or the MCP spec and the learning curve is not justified by the complexity of the use case yet&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What a Production-Ready MCP Implementation Actually Looks Like
&lt;/h2&gt;

&lt;p&gt;Most tutorials show you the happy path. Here is what the production path requires.&lt;/p&gt;

&lt;h3&gt;
  
  
  Server Design: Keep the Surface Area Small
&lt;/h3&gt;

&lt;p&gt;Expose only the tools the agent actually needs for the task. Every tool you expose is part of your attack surface. &lt;/p&gt;

&lt;p&gt;Tool descriptions are read by the LLM treat them as security-sensitive inputs and review them with the same scrutiny you apply to code.&lt;/p&gt;

&lt;h3&gt;
  
  
  Authentication and Authorization: Do Not Trust the Defaults
&lt;/h3&gt;

&lt;p&gt;The MCP spec does not enforce auth you are responsible. Use OAuth with least-privilege scoping per tool, not per server. Rotate credentials.&lt;/p&gt;

&lt;p&gt;Avoid static API keys in config files. If a key leaks, it should grant access to one tool, not your entire integration layer.&lt;/p&gt;

&lt;h3&gt;
  
  
  Observability: Your Forensic Trail
&lt;/h3&gt;

&lt;p&gt;Log every tool call with full context: tool name, parameters, agent identity, session ID, timestamp. Without this, diagnosing a bad agent action after the fact is close to impossible.&lt;/p&gt;

&lt;p&gt;This is not optional infrastructure it is the difference between a system you can debug and one you are flying blind in.&lt;/p&gt;

&lt;h3&gt;
  
  
  Supply Chain Hygiene
&lt;/h3&gt;

&lt;p&gt;Vet every third-party MCP server before connecting it to production. Review source code. Pin versions. Do not auto-update. One malicious package in your agent's tool chain is enough to exfiltrate credentials or compromise infrastructure.&lt;/p&gt;

&lt;p&gt;According to &lt;a href="https://www.mckinsey.com/capabilities/risk-and-resilience/our-insights/cybersecurity/cybersecurity-trends-looking-over-the-horizon" rel="noopener noreferrer"&gt;cybersecurity research&lt;/a&gt;, supply chain risks remain one of the most underestimated attack vectors for organizations adopting new integration standards quickly.&lt;/p&gt;

&lt;h3&gt;
  
  
  Human-in-the-Loop Controls
&lt;/h3&gt;

&lt;p&gt;High-stakes tool calls anything that deletes, writes, deploys, or sends should require explicit human confirmation before execution. Claude Code requires this by default. Your custom agent should too.&lt;/p&gt;

&lt;h2&gt;
  
  
  What It Actually Costs to Adopt MCP
&lt;/h2&gt;

&lt;p&gt;Teams consistently underestimate this. Here is a realistic cost map:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Cost Area&lt;/th&gt;
&lt;th&gt;What Drives It&lt;/th&gt;
&lt;th&gt;Key Insight&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Context token overhead&lt;/td&gt;
&lt;td&gt;Tool schemas loaded at session init&lt;/td&gt;
&lt;td&gt;10 servers × 5 tools = thousands of tokens before work begins&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Server maintenance&lt;/td&gt;
&lt;td&gt;Version pinning, security patches, registry monitoring&lt;/td&gt;
&lt;td&gt;Ongoing ownership, not a one-time setup&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Security implementation&lt;/td&gt;
&lt;td&gt;Auth, audit logging, supply chain vetting&lt;/td&gt;
&lt;td&gt;20-30% budget increase if retrofitted after initial build&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Debugging infrastructure&lt;/td&gt;
&lt;td&gt;Observability tooling, log aggregation, session tracing&lt;/td&gt;
&lt;td&gt;No mature off-the-shelf MCP debugger exists yet&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Engineering ramp-up&lt;/td&gt;
&lt;td&gt;JSON-RPC familiarity, MCP spec learning curve&lt;/td&gt;
&lt;td&gt;1-2 weeks for an experienced backend engineer&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The token overhead is the cost most teams do not model until they see their first production bill. A session loading ten MCP servers before answering a single user query is paying a fixed per-query tax in context tokens. At production query volumes, that accumulates fast.&lt;/p&gt;

&lt;p&gt;The Future: MCP Becomes Invisible Infrastructure&lt;/p&gt;

&lt;p&gt;Here is where this is heading: developers stop asking whether to adopt MCP and start asking how their integration layer is configured.&lt;/p&gt;

&lt;p&gt;The same way you do not debate TCP when you build a web service, you will not debate MCP when you build an agent. The protocol becomes the assumed substrate.&lt;/p&gt;

&lt;p&gt;Google's A2A (Agent-to-Agent) protocol handles agent-to-agent communication. MCP handles agent-to-tool communication. These two standards together define the connective tissue of production agentic AI A2A for agents coordinating with each other, MCP for agents interacting with the world. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025" rel="noopener noreferrer"&gt;AI agent researchs&lt;/a&gt; reinforces this: by 2029, at least half of knowledge workers are expected to be creating, governing, and deploying agents on demand. The integration layer that connects those agents to tools is MCP.&lt;/p&gt;

&lt;p&gt;The teams building security discipline and observability into MCP deployments today are not just solving a current problem. They are building the operational foundation that will matter for the next five years of agentic AI.&lt;/p&gt;

&lt;p&gt;The abstraction will rise. MCP will become invisible. The engineering judgment required to implement it correctly will not.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Model Context Protocol has already won the standards race. 97M monthly SDK downloads, Linux Foundation governance, every major AI vendor committed. The adoption question is closed.&lt;/li&gt;
&lt;li&gt;MCP does not replace REST APIs. It is an orchestration layer above them. Your existing APIs still do the actual work.&lt;/li&gt;
&lt;li&gt;The N×M problem is real. MCP solves it by reducing N×M custom connectors to N+M single-protocol implementations. The integration time savings are measurable and compound as your agent toolset grows.&lt;/li&gt;
&lt;li&gt;Security is the part most tutorials skip and the part that will cost you most. 82% of surveyed implementations have path traversal vulnerabilities. The spec has no built-in auth. You are responsible for all of it.&lt;/li&gt;
&lt;li&gt;Do not add MCP complexity for use cases that one or two direct API calls can handle. The token overhead, server maintenance, and debugging complexity are not free.&lt;/li&gt;
&lt;li&gt;A production-ready MCP deployment requires: minimal tool surface area, proper OAuth, trace-level observability, supply chain vetting, and human-in-the-loop controls on high-stakes actions. Budget for all of it from day one retrofitting security costs 20-30% more than building it in.&lt;/li&gt;
&lt;li&gt;Gartner projects 40% of enterprise apps will embed AI agents by end of 2026. Every one of those agents will need a tool layer. MCP is it.&lt;/li&gt;
&lt;li&gt;Build on MCP when dynamic tool discovery, multi-platform interoperability, or integration scale justifies the complexity. Use direct APIs or function calling when it does not.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Building production MCP systems and have war stories about what actually broke? Drop them in the comments. The useful knowledge is always in the failures nobody writes case studies about.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>machinelearning</category>
      <category>mcp</category>
    </item>
    <item>
      <title>Should You Be Using RAG in 2026?</title>
      <dc:creator>Riddhesh</dc:creator>
      <pubDate>Wed, 15 Apr 2026 09:12:28 +0000</pubDate>
      <link>https://dev.to/riddhesh/should-you-be-using-rag-in-2026-28ef</link>
      <guid>https://dev.to/riddhesh/should-you-be-using-rag-in-2026-28ef</guid>
      <description>&lt;p&gt;Every few months, someone drops a hot take that RAG is dead. Then the next enterprise deal closes, and everyone goes back to building pipelines.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Current Narrative Around RAG
&lt;/h2&gt;

&lt;p&gt;Two positions dominate the conversation right now.&lt;br&gt;
The first: RAG is dead. Long context windows from Anthropic and Google have made retrieval unnecessary. Pass everything into the prompt and call it done.&lt;/p&gt;

&lt;p&gt;The second: RAG is the backbone of enterprise AI and nothing has changed. Stick to the patterns, ship the pipeline.&lt;br&gt;
Both are wrong, and both are reacting to surface-level signals rather than production reality.&lt;/p&gt;

&lt;p&gt;Teams calling RAG dead are confusing context window size with retrieval discipline. Yes, Claude Opus 4.6 now has a 1M-token context window at standard pricing with no long-context surcharge.&lt;/p&gt;

&lt;p&gt;That is a meaningful shift. But a larger context window does not eliminate the need for choosing what goes into it.&lt;/p&gt;

&lt;p&gt;The other camp ignores the real complexity and cost that most teams underestimate until they are already deep in production.&lt;/p&gt;

&lt;p&gt;RAG is not dead. But the version most teams are building in 2026 is not the version that was production-ready two years ago.&lt;/p&gt;

&lt;h2&gt;
  
  
  What RAG Actually Solves
&lt;/h2&gt;

&lt;p&gt;LLMs hallucinate. That is not a bug being patched it is a fundamental property of how autoregressive generation works. Models predict plausible next tokens, not verified facts.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/" rel="noopener noreferrer"&gt;Research&lt;/a&gt; shows that hallucinations remain prevalent in complex reasoning and open-domain factual recall, where error rates can exceed 33%. In a customer-facing application, that is not a product quirk. That is a liability.&lt;/p&gt;

&lt;p&gt;Production-grade RAG, when combined with guardrails and evaluation, reduces hallucinations by 40–96% depending on the stack and the use case, as shown in recent &lt;a href="https://hai.stanford.edu/ai-index" rel="noopener noreferrer"&gt;AI evaluation studies&lt;/a&gt; and benchmarks.&lt;/p&gt;

&lt;p&gt;We have seen this range in our own builds. A well-tuned pipeline with hybrid retrieval and a reranking layer lands closer to the high end of that range. A naive one-shot vector search lands near the low end.&lt;/p&gt;

&lt;p&gt;RAG does not eliminate hallucinations. It constrains the model to a verified knowledge boundary. That distinction matters.&lt;/p&gt;

&lt;h3&gt;
  
  
  Static Knowledge Limitations
&lt;/h3&gt;

&lt;p&gt;Every LLM has a training cutoff. The knowledge is frozen. For most enterprise use cases, policies, product documentation, internal SOPs, and regulatory guidelines, the information that matters is the information that changed last Tuesday, not last year.&lt;/p&gt;

&lt;p&gt;RAG solves this by treating retrieval as a real-time layer. The model never needs to be retrained. Your knowledge base gets updated. The system stays current.&lt;/p&gt;

&lt;h3&gt;
  
  
  Accessing Private and Enterprise Data
&lt;/h3&gt;

&lt;p&gt;This is the most underrated reason enterprises build RAG in 2026. It is not about hallucinations or knowledge cutoffs alone. It is about the fact that the most valuable data a company has is not in any public LLM's training set, and it never will be.&lt;/p&gt;

&lt;p&gt;Customer history, internal research, product specs, compliance documentation. None of that will be absorbed into a foundation model. RAG is how you connect a capable LLM to your actual organizational knowledge without retraining anything.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Market Reality
&lt;/h2&gt;

&lt;p&gt;Here is a number that should end the RAG is dead debate the technology, growing at nearly a 50% CAGR annually, is not dead technology.&lt;/p&gt;

&lt;p&gt;According to &lt;a href="https://www.grandviewresearch.com/industry-analysis/retrieval-augmented-generation-rag-market-report" rel="noopener noreferrer"&gt;Grand View Research&lt;/a&gt;, the global RAG market was valued at USD 1.2 billion in 2024 and is projected to reach USD 11 billion by 2030, growing at a CAGR of 49.1%. &lt;a href="https://www.precedenceresearch.com/retrieval-augmented-generation-market" rel="noopener noreferrer"&gt;Precedence Research&lt;/a&gt; puts the 2026 market at USD 2.76 billion, expanding toward USD 67.42 billion by 2034 at a CAGR of 49.12%.&lt;/p&gt;

&lt;p&gt;Dead tech does not grow this fast.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Metric&lt;/th&gt;
&lt;th&gt;Figure&lt;/th&gt;
&lt;th&gt;Source&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;2024 RAG Market Size&lt;/td&gt;
&lt;td&gt;$1.2 billion&lt;/td&gt;
&lt;td&gt;Grand View Research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;2026 RAG Market Size&lt;/td&gt;
&lt;td&gt;~$2.76 billion&lt;/td&gt;
&lt;td&gt;Precedence Research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;2030 Projected Size&lt;/td&gt;
&lt;td&gt;$11 billion&lt;/td&gt;
&lt;td&gt;Grand View Research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;CAGR (2025–2030)&lt;/td&gt;
&lt;td&gt;~49%&lt;/td&gt;
&lt;td&gt;Grand View Research / Precedence Research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;North America Market Share (2024)&lt;/td&gt;
&lt;td&gt;~37%&lt;/td&gt;
&lt;td&gt;Grand View Research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Asia Pacific Growth Rate&lt;/td&gt;
&lt;td&gt;Fastest-growing region&lt;/td&gt;
&lt;td&gt;Multiple sources&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The demand signal is real. Enterprises in healthcare, legal, finance, and customer service are deploying RAG because the alternative fine-tuning, retraining, or relying on generic model knowledge does not meet compliance, accuracy, or freshness requirements.&lt;/p&gt;

&lt;h2&gt;
  
  
  How RAG Has Evolved (2024 to 2026)
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faeh0egc2xf7wj6rpl340.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faeh0egc2xf7wj6rpl340.png" alt="Evolution of RAG systems from basic retrieval to multi-step pipelines showing improvements in data filtering and retrieval quality" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  RAG v1 vs v2 vs v3
&lt;/h3&gt;

&lt;p&gt;What most engineers built in 2023–2024 was RAG v1. It looked impressive in demos. It fell apart in production. Here is how the architecture has matured:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Generation&lt;/th&gt;
&lt;th&gt;Core Pattern&lt;/th&gt;
&lt;th&gt;Retrieval Method&lt;/th&gt;
&lt;th&gt;Primary Weakness&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;RAG v1 (2023)&lt;/td&gt;
&lt;td&gt;Single vector store + LLM&lt;/td&gt;
&lt;td&gt;Dense embedding search&lt;/td&gt;
&lt;td&gt;Keyword misses, poor ranking&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;RAG v2 (2024)&lt;/td&gt;
&lt;td&gt;Hybrid search + reranking&lt;/td&gt;
&lt;td&gt;Dense + sparse (BM25)&lt;/td&gt;
&lt;td&gt;Chunking quality, latency&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;RAG v3 (2025–2026)&lt;/td&gt;
&lt;td&gt;Agentic pipelines + GraphRAG&lt;/td&gt;
&lt;td&gt;Multi-stage + knowledge graphs&lt;/td&gt;
&lt;td&gt;Complexity, evaluation gaps&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h3&gt;
  
  
  Key Improvements in Modern RAG
&lt;/h3&gt;

&lt;p&gt;Hybrid search is now standard. Dense vector search alone misses exact-match queries. Sparse retrieval (BM25) misses semantic intent. You need both, combined with a fusion layer. Teams that skip this step are leaving retrieval quality on the table.&lt;/p&gt;

&lt;p&gt;Reranking is the layer most teams skip and then regret. A cross-encoder reranker takes the top-k retrieved chunks and reorders them by actual relevance to the query, not just vector similarity. This single step often produces more improvement than spending weeks on embedding model selection.&lt;/p&gt;

&lt;p&gt;Multi-stage pipelines are replacing single-shot retrieval. Query decomposition, sub-query routing, context filtering, and synthesis are now separate stages with their own quality signals. More complex to build. Significantly more reliable in production.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Problems With RAG
&lt;/h2&gt;

&lt;p&gt;RAG works great in demos. It breaks in production.&lt;/p&gt;

&lt;h3&gt;
  
  
  Latency
&lt;/h3&gt;

&lt;p&gt;A naive RAG pipeline adds 800ms to 2 seconds of latency per query. Embedding the user query, hitting the vector store, reranking retrieved chunks, and then generating a response each step has a cost. Users expect sub-500ms responses. You have to engineer aggressively to hit that.&lt;/p&gt;

&lt;p&gt;Caching, pre-fetched indexes, streaming responses, and careful pipeline profiling are not optional. They are the difference between a product and a prototype.&lt;/p&gt;

&lt;h3&gt;
  
  
  Chunking Complexity
&lt;/h3&gt;

&lt;p&gt;Chunking is where most RAG systems silently fail. Fixed-size chunking (splitting documents every 512 tokens) breaks semantic units. A sentence that starts on one chunk and ends on the next becomes unretrievable. An answer that requires context from two adjacent paragraphs gets severed.&lt;/p&gt;

&lt;p&gt;Semantic chunking, section-aware splitting, and metadata-rich indexing are now table stakes, but they add meaningful engineering overhead that nobody budgets for at the start of the project.&lt;/p&gt;

&lt;h3&gt;
  
  
  Retrieval Failures
&lt;/h3&gt;

&lt;p&gt;The model can only work with what you retrieve. If the retrieval step returns three irrelevant chunks, the LLM either hallucinates to fill the gap or produces a useless non-answer. &lt;/p&gt;

&lt;p&gt;Retrieval quality is the ceiling on RAG quality. Most evaluation frameworks miss this because they test generation quality, not retrieval precision.&lt;/p&gt;

&lt;h3&gt;
  
  
  Evaluation Challenges
&lt;/h3&gt;

&lt;p&gt;RAG evaluation is genuinely hard. You need to measure retrieval recall, context relevance, answer faithfulness, and answer correctness four separate signals, each requiring its own methodology. Teams that skip structured evaluation ship systems they cannot diagnose when they degrade.&lt;/p&gt;

&lt;h2&gt;
  
  
  RAG vs Fine-Tuning vs Agents
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F39k11akxnadbgaoswyt1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F39k11akxnadbgaoswyt1.png" alt="RAG vs Fine-Tuning vs Agents comparison diagram showing differences in architecture" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is the question every technical founder should have a clear answer to before choosing an architecture.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Approach&lt;/th&gt;
&lt;th&gt;Pros&lt;/th&gt;
&lt;th&gt;Cons&lt;/th&gt;
&lt;th&gt;Best Use Case&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;RAG&lt;/td&gt;
&lt;td&gt;No retraining, fresh knowledge, explainable, cost-effective&lt;/td&gt;
&lt;td&gt;Latency overhead, retrieval failures, chunking complexity&lt;/td&gt;
&lt;td&gt;Dynamic knowledge, private data, compliance-sensitive use cases&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Fine-Tuning&lt;/td&gt;
&lt;td&gt;Faster inference, domain-specific tone/format, no retrieval&lt;/td&gt;
&lt;td&gt;Expensive, knowledge is static, hard to update, requires quality data&lt;/td&gt;
&lt;td&gt;Consistent style/format, specialized reasoning, narrow domains&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Agents&lt;/td&gt;
&lt;td&gt;Dynamic, multi-step, tool-using, self-correcting&lt;/td&gt;
&lt;td&gt;High latency, unpredictable, expensive per query, hard to evaluate&lt;/td&gt;
&lt;td&gt;Complex workflows, autonomous research, multi-system orchestration&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;RAG wins when your primary constraint is knowledge freshness and accuracy over private data. Fine-tuning wins when your primary constraint is style consistency, inference speed, or domain-specific formatting. Agents win when the task is inherently multi-step and requires dynamic decision-making.&lt;/p&gt;

&lt;p&gt;The mistake most teams make is defaulting to one architecture for everything. The right call is almost always a hybrid: RAG for grounding, light fine-tuning for tone consistency, and agentic orchestration for workflows that need planning.&lt;/p&gt;

&lt;p&gt;Why are agents rising so fast? Because modern LLMs are capable enough to execute multi-step plans reliably. RAG becomes one tool that an agent calls, rather than the whole system. This is the direction production architectures are moving.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Modern AI Companies Are Changing
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Long Context Windows Are Real But Not a Replacement
&lt;/h3&gt;

&lt;p&gt;Anthropic made its 1M-token context window generally available for Claude Opus 4.6 and Sonnet 4.6, removing the long-context pricing premium entirely. A 900,000-token request now costs the same per token as a 9,000-token one.&lt;/p&gt;

&lt;p&gt;Claude Opus 4.6 scored 78.3% on MRCR v2 at 1M tokens, the highest recall among frontier models compared to Gemini 3 Pro's 26.3% at the same context length, according to &lt;a href="https://www.anthropic.com/news" rel="noopener noreferrer"&gt;official reports&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;This matters. It changes the cost math for some use cases. A team that previously used RAG to process a 200-page legal document can now consider stuffing the full document into a single prompt for a fraction of what it cost six months ago.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why Context Does Not Equal Retrieval
&lt;/h3&gt;

&lt;p&gt;But here is what gets lost in the benchmarks for example 1M token context window and a well-structured retrieval pipeline solve different problems.&lt;/p&gt;

&lt;p&gt;Context windows are static per-request. You load what you know needs to be there. RAG is dynamic it retrieves what is relevant at query time across a corpus that may contain millions of documents.&lt;/p&gt;

&lt;p&gt;No context window is large enough to hold an enterprise knowledge base. And even if it were, you would not want to pay to process it in full on every single query.&lt;/p&gt;

&lt;p&gt;As Anthropic's own pricing analysis shows, flat-rate long-context pricing makes the economics more predictable and linear but a single 1M-token prompt costs $3 at Sonnet rates, which adds up significantly at production query volumes.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where RAG Still Fits
&lt;/h3&gt;

&lt;p&gt;RAG remains the right call for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Knowledge bases larger than any context window&lt;/li&gt;
&lt;li&gt;High-query-volume workloads where cost per query matters&lt;/li&gt;
&lt;li&gt;Use cases requiring source citations and explainability&lt;/li&gt;
&lt;li&gt;Real-time document ingestion where the corpus updates continuously&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Long context is a tool. RAG is an architecture. They are increasingly complementary, not competitive.&lt;/p&gt;

&lt;h2&gt;
  
  
  When You SHOULD Use RAG
&lt;/h2&gt;

&lt;p&gt;Use RAG when all of the following are true:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;RAG makes sense when your data is private, proprietary, or constantly evolving because pretrained models can’t keep up with changing information, while retrieval ensures responses stay aligned with your latest data.&lt;/li&gt;
&lt;li&gt;It becomes essential in scenarios where accuracy and citation matter since responses can be tied back to actual source documents, making them easier to trust, verify, and audit.&lt;/li&gt;
&lt;li&gt;For large knowledge bases that exceed practical context window limits, retrieval helps narrow things down only the most relevant chunks are injected, which improves both efficiency and output quality.&lt;/li&gt;
&lt;li&gt;In regulated industries like healthcare, legal, or finance, having explainable outputs is critical RAG provides a clear link between responses and approved knowledge sources.&lt;/li&gt;
&lt;li&gt;When answers need to come from specific internal documents rather than general model knowledge, retrieval acts as the bridge between your data and the model’s reasoning.&lt;/li&gt;
&lt;li&gt;If your information changes frequently, maintaining a vector database is far more practical than retraining or fine-tuning models every time something updates.&lt;/li&gt;
&lt;li&gt;That said, all of this only works well if the system is implemented properly retrieval quality depends heavily on chunking, embeddings, and ranking, so weak setups can actually hurt performance.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Enterprise data indicates that organizations are choosing RAG for 30-60% of their AI use cases specifically where high accuracy, transparency, and reliable outputs over proprietary data are required.&lt;/p&gt;

&lt;h2&gt;
  
  
  When You SHOULD NOT Use RAG
&lt;/h2&gt;

&lt;p&gt;Stop reaching for RAG when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Tasks that rely on general knowledge like coding help, writing, or basic Q&amp;amp;A adding retrieval usually creates unnecessary overhead without improving results.&lt;/li&gt;
&lt;li&gt;Smaller, static datasets don’t really justify the complexity if everything fits into a prompt, direct context injection is often cleaner and more reliable.&lt;/li&gt;
&lt;li&gt;In low-query environments, the trade-off shifts passing full documents per request can be simpler than maintaining a full retrieval pipeline.&lt;/li&gt;
&lt;li&gt;Real-time or latency-sensitive applications can struggle with RAG, since the added retrieval and reranking steps introduce delays that may not be acceptable.&lt;/li&gt;
&lt;li&gt;Teams without experience in retrieval systems often run into issues poor chunking or irrelevant results can degrade output quality instead of improving it.&lt;/li&gt;
&lt;li&gt;During early prototyping, simplicity wins validating the idea with basic prompting is faster, and retrieval can always be layered in once the need becomes clear.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Many teams would be better served by a well-crafted system prompt with a few embedded documents than by a full RAG pipeline. Start simple. Add retrieval when you actually need it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What a Production-Ready RAG Stack Looks Like in 2026
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frpd419qs8h7k0q31ilet.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frpd419qs8h7k0q31ilet.png" alt="Production RAG architecture diagram showing user query, retrieval, reranking, LLM processing, and response flow with supporting layers like caching and observability" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Retrieval Layer
&lt;/h3&gt;

&lt;p&gt;Hybrid search combining dense vector search and BM25 sparse retrieval with a score fusion layer. &lt;a href="https://code-b.dev/blog/llm-embeddings" rel="noopener noreferrer"&gt;Embedding&lt;/a&gt; model selection matters domain-specific embeddings consistently outperform general-purpose ones for specialized corpora, which is why understanding embeddings before you pick a model is worth the time.&lt;/p&gt;

&lt;h3&gt;
  
  
  Reranking
&lt;/h3&gt;

&lt;p&gt;A cross-encoder reranker sits between retrieval and generation. It takes the top-20 retrieved chunks and reorders them by actual query relevance before passing the top-5 to the LLM. This is the highest-leverage improvement most teams are not making.&lt;/p&gt;

&lt;h3&gt;
  
  
  Context Filtering
&lt;/h3&gt;

&lt;p&gt;Metadata filters, access controls, and document freshness scoring. Not every retrieved document should reach the model. &lt;/p&gt;

&lt;p&gt;Stale documents, low-confidence chunks, and access-restricted content need to be filtered before they can contaminate the response.&lt;/p&gt;

&lt;h3&gt;
  
  
  Caching
&lt;/h3&gt;

&lt;p&gt;Semantic caching for repeated or similar queries. Query-level caching reduces both latency and cost at scale. &lt;/p&gt;

&lt;p&gt;Anthropic's prompt caching makes long-context RAG workflows significantly cheaper for repeated queries against the same document corpus the first read costs full price, and subsequent reads come at a fraction of the cost.&lt;/p&gt;

&lt;h3&gt;
  
  
  Observability
&lt;/h3&gt;

&lt;p&gt;This is the piece that kills RAG systems in production. You need trace-level logging across every pipeline stage and what was retrieved, what was scored, what was passed to the model, and what was generated. &lt;/p&gt;

&lt;p&gt;Without this, debugging a bad answer is impossible. Tooling like LangSmith, Arize, and custom logging layers are no longer optional.&lt;/p&gt;

&lt;h2&gt;
  
  
  What It Actually Costs to Run RAG
&lt;/h2&gt;

&lt;p&gt;Teams consistently underestimate this. Here is a realistic breakdown:&lt;/p&gt;

&lt;h3&gt;
  
  
  Infrastructure cost
&lt;/h3&gt;

&lt;p&gt;Vector database, embedding model inference, and reranking model inference. At modest scale (100K queries/month), this is $500-$2,000/month in infrastructure depending on your stack. At 1M+ queries/month, this becomes a line item in the P&amp;amp;L.&lt;/p&gt;

&lt;h3&gt;
  
  
  LLM cost
&lt;/h3&gt;

&lt;p&gt;At Sonnet 4.6 pricing ($3 input / $15 output per million tokens), a RAG response that passes 5 chunks of 400 tokens each costs roughly $0.006 per query in input tokens alone. At 100K queries/month, that is $600/month just in context passing before generation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Latency tradeoff
&lt;/h3&gt;

&lt;p&gt;Production RAG systems average 1.2-2.5 seconds end-to-end. If you need sub-500ms, you are engineering against the grain of the architecture. Aggressive caching and streaming can help, but they add complexity.&lt;/p&gt;

&lt;h3&gt;
  
  
  Maintenance complexity
&lt;/h3&gt;

&lt;p&gt;Someone has to own the chunking pipeline, the embedding pipeline, the index refresh schedule, and the retrieval evaluation framework. This is not a set-and-forget system. Budget accordingly.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Area&lt;/th&gt;
&lt;th&gt;What Drives Cost / Complexity&lt;/th&gt;
&lt;th&gt;Key Insight&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Infrastructure&lt;/td&gt;
&lt;td&gt;Vector DB, embeddings, reranking&lt;/td&gt;
&lt;td&gt;Costs scale non-linearly as query volume increases&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Token Consumption&lt;/td&gt;
&lt;td&gt;Retrieved context size&lt;/td&gt;
&lt;td&gt;More chunks = better accuracy, but higher per-query cost&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Generation&lt;/td&gt;
&lt;td&gt;Output length + model choice&lt;/td&gt;
&lt;td&gt;Often overlooked, but can exceed retrieval costs&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Latency&lt;/td&gt;
&lt;td&gt;Retrieval + reranking pipeline&lt;/td&gt;
&lt;td&gt;Speed trade-offs are architectural, not easily optimized&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;System Maintenance&lt;/td&gt;
&lt;td&gt;Pipelines, indexing, evaluation&lt;/td&gt;
&lt;td&gt;Requires continuous ownership, not a one-time setup&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;RAG doesn’t just add cost it shifts your system from a simple API call to a continuously managed infrastructure layer.&lt;/p&gt;

&lt;h2&gt;
  
  
  How RAG Is Becoming Invisible Infrastructure in AI Systems
&lt;/h2&gt;

&lt;p&gt;Here is where I think this is heading: RAG stops being a product decision and starts being infrastructure.&lt;/p&gt;

&lt;p&gt;In the same way that most teams do not think about how their application talks to a database it just does RAG will become an invisible layer in AI systems. The retrieval, chunking, and indexing will happen automatically as part of the model's operating environment.&lt;/p&gt;

&lt;p&gt;Anthropic's push toward MCP (Model Context Protocol) is an early signal of this. As of early 2026, MCP has been adopted by OpenAI, Google, and Microsoft, and donated to the Linux Foundation becoming an industry standard for AI agent integration with external data sources.&lt;/p&gt;

&lt;p&gt;The question will not be "should we use RAG?" but "how is our context layer configured?" The abstraction will rise. The engineers building the abstraction will still need to understand everything in this post.&lt;/p&gt;

&lt;p&gt;RAG is not going away. It is going deeper.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;RAG is not dead. The market is growing at ~49% CAGR. Dead tech does not scale like that.&lt;/li&gt;
&lt;li&gt;Long context windows from Anthropic and Google change the cost math but do not eliminate the need for retrieval at enterprise scale.&lt;/li&gt;
&lt;li&gt;The version of RAG that fails in production is RAG v1. Hybrid search, reranking, and observability are now required, not optional.&lt;/li&gt;
&lt;li&gt;RAG reduces hallucinations by 40–96% in well-engineered stacks. But retrieval quality is the ceiling. You cannot generate your way out of a bad retrieval.&lt;/li&gt;
&lt;li&gt;The real competition is not RAG vs long context. It is RAG vs fine-tuning vs agents and the right answer is almost always a hybrid depending on the use case.&lt;/li&gt;
&lt;li&gt;Do not build RAG because it is the trendy pattern. Build it when your use case genuinely requires dynamic retrieval over a large, private, frequently-updated knowledge corpus.&lt;/li&gt;
&lt;li&gt;Budget for the full stack embedding pipelines, vector databases, reranking, caching, and critically observability. RAG is not a plug-and-play feature. It is a system.&lt;/li&gt;
&lt;li&gt;RAG is becoming infrastructure. The teams that build operational discipline around it now will be positioned well as the abstraction layer rises.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you are building production AI systems and want to share what is actually working in your RAG stack, drop it in the comments. The interesting decisions are always in the engineering details no one writes blog posts about.&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>programming</category>
      <category>llm</category>
    </item>
    <item>
      <title>CI/CD Pipeline for React Native</title>
      <dc:creator>Riddhesh</dc:creator>
      <pubDate>Fri, 15 Mar 2024 12:15:06 +0000</pubDate>
      <link>https://dev.to/riddhesh/cicd-pipeline-for-react-native-133f</link>
      <guid>https://dev.to/riddhesh/cicd-pipeline-for-react-native-133f</guid>
      <description>&lt;p&gt;&lt;a href="https://code-b.dev/blog/cicd-for-react-native" rel="noopener noreferrer"&gt;https://code-b.dev/blog/cicd-for-react-native&lt;/a&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>reactnative</category>
      <category>react</category>
      <category>beginners</category>
    </item>
    <item>
      <title>Use hooks to transform your react native projects</title>
      <dc:creator>Riddhesh</dc:creator>
      <pubDate>Wed, 10 Jan 2024 07:46:21 +0000</pubDate>
      <link>https://dev.to/riddhesh/use-hooks-to-transform-your-react-native-projects-5cbl</link>
      <guid>https://dev.to/riddhesh/use-hooks-to-transform-your-react-native-projects-5cbl</guid>
      <description>&lt;p&gt;React Native, since its introduction, has transformed and improved mobile app development and performance through features like hooks.&lt;br&gt;
Hooks enable access to React state and lifecycle features in function components, overcoming the limitations of class components. They offer benefits such as enhanced code reusability, improved readability, smaller component sizes, easier state management, and improved testing. &lt;br&gt;
Key React Native hooks include &lt;code&gt;useState&lt;/code&gt; for managing component-specific states, &lt;code&gt;useEffect&lt;/code&gt; for handling side effects, &lt;code&gt;useContext&lt;/code&gt; for accessing context in functional components, and custom hooks for promoting code reusability by encapsulating common state logic. &lt;br&gt;
These hooks have revolutionized React Native development by simplifying and optimizing the coding process. &lt;br&gt;
Here's an example of how to use the &lt;code&gt;useState&lt;/code&gt; hook in a functional component:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;br&gt;
import React, { useState } from 'react';&lt;br&gt;
import { View, Text, Button } from 'react-native';&lt;br&gt;
function Counter() {&lt;br&gt;
const [count, setCount] = useState(0);&lt;br&gt;
return (&lt;br&gt;
&amp;lt;View&amp;gt;&lt;br&gt;
&amp;lt;Text&amp;gt;{count}&amp;lt;/Text&amp;gt;&lt;br&gt;
&amp;lt;Button title="Increment" onPress={() =&amp;gt; setCount(count + 1)} /&amp;gt;&lt;br&gt;
&amp;lt;/View&amp;gt;&lt;br&gt;
);&lt;br&gt;
}&lt;br&gt;
export default Counter;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;P.S.&lt;br&gt;
Felt like doing an informative post as a first post, this is based on a more detailed blog post I made on my website, I don't know if I am allowed to share the link here but anyway, it's &lt;a href="https://code-b.dev/blog/react-native-hooks" rel="noopener noreferrer"&gt;this&lt;/a&gt;&lt;/p&gt;

</description>
      <category>react</category>
      <category>reactnative</category>
      <category>firstpost</category>
      <category>tutorial</category>
    </item>
  </channel>
</rss>
