<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Prateek Chaudhary</title>
    <description>The latest articles on DEV Community by Prateek Chaudhary (@prateekdalal).</description>
    <link>https://dev.to/prateekdalal</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3758371%2F4138c958-6921-4ed0-ba53-bb39b6748925.jpeg</url>
      <title>DEV Community: Prateek Chaudhary</title>
      <link>https://dev.to/prateekdalal</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/prateekdalal"/>
    <language>en</language>
    <item>
      <title>Kakveda v1.0.3 – One-Line AI Agent Governance Integration</title>
      <dc:creator>Prateek Chaudhary</dc:creator>
      <pubDate>Tue, 17 Feb 2026 14:01:17 +0000</pubDate>
      <link>https://dev.to/prateekdalal/kakveda-v103-one-line-ai-agent-governance-integration-2al5</link>
      <guid>https://dev.to/prateekdalal/kakveda-v103-one-line-ai-agent-governance-integration-2al5</guid>
      <description>&lt;p&gt;Just released Kakveda v1.0.3.&lt;/p&gt;

&lt;p&gt;Big change: SDK-first integration.&lt;/p&gt;

&lt;p&gt;Before:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Manual warn calls&lt;/li&gt;
&lt;li&gt;Manual event publishing&lt;/li&gt;
&lt;li&gt;Separate registration scripts&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now:&lt;br&gt;
from kakveda_sdk import KakvedaAgent&lt;/p&gt;

&lt;p&gt;agent = KakvedaAgent()&lt;/p&gt;

&lt;p&gt;agent.execute(&lt;br&gt;
    prompt="export data",&lt;br&gt;
    tool_name="data_exporter",&lt;br&gt;
    execute_fn=my_func&lt;br&gt;
)&lt;/p&gt;

&lt;p&gt;That’s it.&lt;/p&gt;

&lt;p&gt;SDK handles:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Preflight policy checks&lt;/li&gt;
&lt;li&gt;Trace ingestion&lt;/li&gt;
&lt;li&gt;Dashboard registration&lt;/li&gt;
&lt;li&gt;Heartbeat&lt;/li&gt;
&lt;li&gt;Retry &amp;amp; circuit breaker&lt;/li&gt;
&lt;li&gt;Strict mode enforcement
Removed legacy helpers to simplify developer experience.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you're building AI agents in production, especially multi-agent systems, feedback is welcome.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>python</category>
      <category>showdev</category>
    </item>
    <item>
      <title>Building Failure Intelligence for AI Agents</title>
      <dc:creator>Prateek Chaudhary</dc:creator>
      <pubDate>Mon, 16 Feb 2026 16:04:11 +0000</pubDate>
      <link>https://dev.to/prateekdalal/building-failure-intelligence-for-ai-agents-170d</link>
      <guid>https://dev.to/prateekdalal/building-failure-intelligence-for-ai-agents-170d</guid>
      <description>&lt;p&gt;When you run AI agents in production, you quickly realize:&lt;/p&gt;

&lt;p&gt;The dangerous failures aren’t random.&lt;br&gt;
They’re recurring patterns.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Similar hallucination structures&lt;/li&gt;
&lt;li&gt;Repeated tool-call mistakes&lt;/li&gt;
&lt;li&gt;Prompt injection variants&lt;/li&gt;
&lt;li&gt;Context leakage patterns&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Most tools give you logs.&lt;br&gt;
Some give you tracing.&lt;br&gt;
Few give you structured failure memory.&lt;/p&gt;

&lt;p&gt;I’ve been exploring a model where:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Every failure becomes a canonical entity&lt;/li&gt;
&lt;li&gt;A deterministic fingerprint is generated for executions&lt;/li&gt;
&lt;li&gt;New executions are matched against historical failures&lt;/li&gt;
&lt;li&gt;A policy engine maps confidence → allow / warn / block&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The key idea:&lt;/p&gt;

&lt;p&gt;Don’t modify the LLM.&lt;br&gt;
Don’t rely only on prompts.&lt;br&gt;
Insert a deterministic governance layer before execution.&lt;/p&gt;

&lt;p&gt;This turns failure history into enforcement intelligence.&lt;/p&gt;

&lt;p&gt;Still early, but curious:&lt;br&gt;
LINK : &lt;a href="https://github.com/prateekdevisingh/kakveda" rel="noopener noreferrer"&gt;https://github.com/prateekdevisingh/kakveda&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;How are others handling repeat failure patterns in agent-based systems?&lt;/p&gt;

&lt;h1&gt;
  
  
  opensource #llm #agents #devops #aigovernance
&lt;/h1&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>architecture</category>
      <category>llm</category>
    </item>
    <item>
      <title>Operating AI in Production Is an Ops Problem</title>
      <dc:creator>Prateek Chaudhary</dc:creator>
      <pubDate>Sat, 07 Feb 2026 11:32:34 +0000</pubDate>
      <link>https://dev.to/prateekdalal/operating-ai-in-production-is-an-ops-problem-2io6</link>
      <guid>https://dev.to/prateekdalal/operating-ai-in-production-is-an-ops-problem-2io6</guid>
      <description>&lt;p&gt;Over the last year, I’ve been working with LLMs and AI systems that actually run in production — not demos, not notebooks, not proof-of-concepts.&lt;/p&gt;

&lt;p&gt;What surprised me most wasn’t model behavior.&lt;br&gt;
It was how quickly operational assumptions broke.&lt;/p&gt;

&lt;p&gt;From an ops and platform perspective, AI systems don’t fail like models.&lt;br&gt;
They fail like systems.&lt;/p&gt;

&lt;p&gt;What breaks first in real environments&lt;/p&gt;

&lt;p&gt;When AI systems move into production, the early issues are rarely about accuracy.&lt;/p&gt;

&lt;p&gt;Instead, teams struggle with:&lt;/p&gt;

&lt;p&gt;unclear decision boundaries&lt;/p&gt;

&lt;p&gt;non-reproducible behavior&lt;/p&gt;

&lt;p&gt;missing audit trails&lt;/p&gt;

&lt;p&gt;no safe rollback paths&lt;/p&gt;

&lt;p&gt;uncomfortable “why did this happen?” questions&lt;/p&gt;

&lt;p&gt;Most existing tooling focuses on observing outputs.&lt;br&gt;
Very little focuses on governing behavior.&lt;/p&gt;

&lt;p&gt;Observability helps, but it’s reactive&lt;/p&gt;

&lt;p&gt;We already know how to observe software:&lt;/p&gt;

&lt;p&gt;logs&lt;/p&gt;

&lt;p&gt;metrics&lt;/p&gt;

&lt;p&gt;traces&lt;/p&gt;

&lt;p&gt;alerts&lt;/p&gt;

&lt;p&gt;AI observability tools extend this to:&lt;/p&gt;

&lt;p&gt;drift&lt;/p&gt;

&lt;p&gt;cost&lt;/p&gt;

&lt;p&gt;latency&lt;/p&gt;

&lt;p&gt;token usage&lt;/p&gt;

&lt;p&gt;All useful — but mostly after the fact.&lt;/p&gt;

&lt;p&gt;In production systems, knowing what happened is not enough.&lt;br&gt;
You also need to know:&lt;/p&gt;

&lt;p&gt;whether it should have happened&lt;/p&gt;

&lt;p&gt;whether it should happen again&lt;/p&gt;

&lt;p&gt;whether it should be allowed at all&lt;/p&gt;

&lt;p&gt;The core mismatch&lt;/p&gt;

&lt;p&gt;LLMs reason probabilistically.&lt;br&gt;
Production systems expect determinism.&lt;/p&gt;

&lt;p&gt;Trying to force AI to behave like traditional software doesn’t work.&lt;br&gt;
But letting AI directly execute decisions inside deterministic systems also doesn’t work.&lt;/p&gt;

&lt;p&gt;So we started experimenting with a different boundary:&lt;/p&gt;

&lt;p&gt;AI can reason.&lt;br&gt;
Deterministic systems decide.&lt;br&gt;
Execution must remain controlled.&lt;/p&gt;

&lt;p&gt;Separating reasoning from execution&lt;/p&gt;

&lt;p&gt;Once you separate these concerns, a lot of things become clearer:&lt;/p&gt;

&lt;p&gt;AI suggestions can be evaluated before execution&lt;/p&gt;

&lt;p&gt;policies can block or correct unsafe actions&lt;/p&gt;

&lt;p&gt;failures become structured signals, not surprises&lt;/p&gt;

&lt;p&gt;accountability boundaries become explicit&lt;/p&gt;

&lt;p&gt;This is a familiar pattern in ops — just applied to intelligence.&lt;/p&gt;

&lt;p&gt;Why I started working on Kakveda&lt;/p&gt;

&lt;p&gt;This line of thinking led me to start working on Kakveda, an open-source project focused on intelligence monitoring, observability, and deterministic control for AI systems.&lt;/p&gt;

&lt;p&gt;The goal isn’t to replace models or agents.&lt;br&gt;
It’s to supervise them.&lt;/p&gt;

&lt;p&gt;Kakveda sits around AI systems and focuses on:&lt;/p&gt;

&lt;p&gt;observing how AI behaves over time&lt;/p&gt;

&lt;p&gt;enforcing rules before actions execute&lt;/p&gt;

&lt;p&gt;capturing failures as first-class events&lt;/p&gt;

&lt;p&gt;keeping execution predictable&lt;/p&gt;

&lt;p&gt;In short: making AI systems operable.&lt;/p&gt;

&lt;p&gt;What Kakveda is not&lt;/p&gt;

&lt;p&gt;To be clear, Kakveda is not:&lt;/p&gt;

&lt;p&gt;a prompt framework&lt;/p&gt;

&lt;p&gt;an agent toolkit&lt;/p&gt;

&lt;p&gt;an LLM wrapper&lt;/p&gt;

&lt;p&gt;a chatbot platform&lt;/p&gt;

&lt;p&gt;It doesn’t try to make AI smarter.&lt;br&gt;
It tries to make AI safer to run.&lt;/p&gt;

&lt;p&gt;Why open source&lt;/p&gt;

&lt;p&gt;Governance and control layers should not be opaque.&lt;/p&gt;

&lt;p&gt;If AI already introduces uncertainty, the systems supervising it should be:&lt;/p&gt;

&lt;p&gt;inspectable&lt;/p&gt;

&lt;p&gt;auditable&lt;/p&gt;

&lt;p&gt;adaptable&lt;/p&gt;

&lt;p&gt;Open source allows this to evolve based on real failures, not theoretical design.&lt;/p&gt;

&lt;p&gt;Kakveda is early-stage and opinionated — and that’s intentional.&lt;/p&gt;

&lt;p&gt;The bigger takeaway&lt;/p&gt;

&lt;p&gt;As AI adoption grows, the most important question won’t be:&lt;/p&gt;

&lt;p&gt;“How powerful is this model?”&lt;/p&gt;

&lt;p&gt;It will be:&lt;/p&gt;

&lt;p&gt;“Do we understand and control what this system is allowed to do?”&lt;/p&gt;

&lt;p&gt;That’s an ops question.&lt;br&gt;
And ops questions deserve first-class systems.&lt;/p&gt;

&lt;p&gt;If you’re operating AI systems in production — especially from a DevOps, SRE, or platform perspective — I’d love to hear what’s breaking for you.&lt;/p&gt;

</description>
      <category>devops</category>
      <category>ai</category>
      <category>opensource</category>
      <category>observability</category>
    </item>
  </channel>
</rss>
