<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Siddhartha Reddy</title>
    <description>The latest articles on DEV Community by Siddhartha Reddy (@siddhartha_reddy).</description>
    <link>https://dev.to/siddhartha_reddy</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3748507%2F6971bcb7-34df-4b38-8665-66909123139c.jpg</url>
      <title>DEV Community: Siddhartha Reddy</title>
      <link>https://dev.to/siddhartha_reddy</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/siddhartha_reddy"/>
    <language>en</language>
    <item>
      <title>Choosing the Right Model (Not the Best One)</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Tue, 28 Apr 2026 05:56:34 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/choosing-the-right-model-not-the-best-one-1f4f</link>
      <guid>https://dev.to/siddhartha_reddy/choosing-the-right-model-not-the-best-one-1f4f</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;The best model is rarely the right model.&lt;br&gt;&lt;br&gt;
And chasing it is one of the biggest mistakes in AI.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Default Thinking
&lt;/h2&gt;

&lt;p&gt;Most people ask:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“What’s the best model for this problem?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;So they:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Look for highest accuracy
&lt;/li&gt;
&lt;li&gt;Pick the most advanced architecture
&lt;/li&gt;
&lt;li&gt;Optimize benchmarks
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This works in research.&lt;br&gt;&lt;br&gt;
👉 It fails in production.&lt;/p&gt;


&lt;h2&gt;
  
  
  🧠 The Real Question
&lt;/h2&gt;

&lt;p&gt;Instead of asking:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“What’s the best model?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Ask:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;“What’s the right model for this system?”&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;


&lt;h2&gt;
  
  
  ⚙️ Why “Best” Doesn’t Work in Real Systems
&lt;/h2&gt;

&lt;p&gt;Real-world systems have constraints:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Latency (how fast it responds)
&lt;/li&gt;
&lt;li&gt;Cost (compute + infra)
&lt;/li&gt;
&lt;li&gt;Scale (number of users)
&lt;/li&gt;
&lt;li&gt;Reliability (consistency)
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The “best” model often violates these.&lt;/p&gt;


&lt;h2&gt;
  
  
  ⚖️ The Tradeoff Triangle
&lt;/h2&gt;

&lt;p&gt;Every model choice is a tradeoff between:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Accuracy ↔ Latency ↔ Cost
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You can’t maximize all three.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;High accuracy → slower, expensive
&lt;/li&gt;
&lt;li&gt;Low latency → simpler models
&lt;/li&gt;
&lt;li&gt;Low cost → compromises elsewhere
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Choosing a model = choosing a tradeoff&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 Example (Real-World Scenario)
&lt;/h2&gt;

&lt;p&gt;Imagine a recommendation system:&lt;/p&gt;

&lt;h3&gt;
  
  
  Option A:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Complex deep learning model
&lt;/li&gt;
&lt;li&gt;High accuracy
&lt;/li&gt;
&lt;li&gt;Slow response
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Option B:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Simpler model
&lt;/li&gt;
&lt;li&gt;Slightly lower accuracy
&lt;/li&gt;
&lt;li&gt;Fast + cheap
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 In production, Option B often wins.&lt;/p&gt;

&lt;p&gt;Why?&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Because users care about speed and consistency.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  ⚠️ Overengineering is a Real Problem
&lt;/h2&gt;

&lt;p&gt;Many systems fail because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Model is too complex
&lt;/li&gt;
&lt;li&gt;Hard to deploy
&lt;/li&gt;
&lt;li&gt;Hard to debug
&lt;/li&gt;
&lt;li&gt;Expensive to scale
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Complexity increases cost without proportional value.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧱 Fit the Model to the System
&lt;/h2&gt;

&lt;p&gt;Your model should match:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Use case
&lt;/h3&gt;

&lt;p&gt;Real-time vs batch&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Data availability
&lt;/h3&gt;

&lt;p&gt;Small vs large datasets&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Infrastructure
&lt;/h3&gt;

&lt;p&gt;What you can actually run&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Business goals
&lt;/h3&gt;

&lt;p&gt;Speed vs accuracy vs cost&lt;/p&gt;




&lt;h2&gt;
  
  
  🔄 Iteration Matters More Than Perfection
&lt;/h2&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Building the “perfect” model
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Focus on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Deploying a working model
&lt;/li&gt;
&lt;li&gt;Measuring performance
&lt;/li&gt;
&lt;li&gt;Improving iteratively
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Speed of iteration &amp;gt; initial perfection&lt;/p&gt;




&lt;h2&gt;
  
  
  🚀 What Actually Works
&lt;/h2&gt;

&lt;p&gt;Start simple:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Baseline model
&lt;/li&gt;
&lt;li&gt;Measure
&lt;/li&gt;
&lt;li&gt;Improve
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Increase complexity only if needed
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧠 Key Insight
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;A slightly worse model in a strong system&lt;br&gt;&lt;br&gt;
beats&lt;br&gt;&lt;br&gt;
A perfect model in a weak system&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;AI systems don’t succeed because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;They use the best models
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They succeed because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;They use the right models for the system&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Don’t chase the best model.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Choose the model that fits your constraints.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Anyone can benchmark models.&lt;/p&gt;

&lt;p&gt;Very few can:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Choose models that actually work in production&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s where real engineering happens.&lt;/p&gt;




</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>systemdesign</category>
      <category>mlops</category>
    </item>
    <item>
      <title>Data First, Model Later: The Right Way to Build AI Systems</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Thu, 23 Apr 2026 11:12:39 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/data-first-model-later-the-right-way-to-build-ai-systems-2l2g</link>
      <guid>https://dev.to/siddhartha_reddy/data-first-model-later-the-right-way-to-build-ai-systems-2l2g</guid>
      <description>&lt;h2&gt;
  
  
  Most AI systems fail not because of bad models, but because of bad data. Here’s why data should come first.
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Most AI systems don’t fail because of bad models.&lt;br&gt;&lt;br&gt;
They fail because of bad data.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Common Mistake
&lt;/h2&gt;

&lt;p&gt;Most teams start like this:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Choose a model
&lt;/li&gt;
&lt;li&gt;Train it
&lt;/li&gt;
&lt;li&gt;Then figure out the data
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This is backwards.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧠 The Reality
&lt;/h2&gt;

&lt;p&gt;Models don’t create intelligence.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Data does.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The model just:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Learns patterns
&lt;/li&gt;
&lt;li&gt;From the data you give it
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your data is:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Incomplete
&lt;/li&gt;
&lt;li&gt;Noisy
&lt;/li&gt;
&lt;li&gt;Misaligned
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Your system will fail no matter how good the model is.&lt;/p&gt;




&lt;h2&gt;
  
  
  📊 Why Data Matters More Than Models
&lt;/h2&gt;

&lt;p&gt;A simple rule:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Better data + simple model&lt;br&gt;&lt;br&gt;
beats&lt;br&gt;&lt;br&gt;
Bad data + complex model&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧩 What “Good Data” Actually Means
&lt;/h2&gt;

&lt;p&gt;Not just:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Large datasets
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Relevant
&lt;/h3&gt;

&lt;p&gt;Matches real-world use cases&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Clean
&lt;/h3&gt;

&lt;p&gt;Minimal errors and inconsistencies&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Representative
&lt;/h3&gt;

&lt;p&gt;Covers actual production scenarios&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Updated
&lt;/h3&gt;

&lt;p&gt;Reflects current patterns (not outdated)&lt;/p&gt;




&lt;h2&gt;
  
  
  ⚠️ The Biggest Problem: Training ≠ Production Data
&lt;/h2&gt;

&lt;p&gt;In training:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean datasets
&lt;/li&gt;
&lt;li&gt;Structured inputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In production:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Missing values
&lt;/li&gt;
&lt;li&gt;Noise
&lt;/li&gt;
&lt;li&gt;Unexpected formats
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This mismatch is where systems break.&lt;/p&gt;




&lt;h2&gt;
  
  
  🔄 Data is Not Static
&lt;/h2&gt;

&lt;p&gt;Most people think:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Collect data → Train → Done
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Reality:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Collect → Clean → Use → Monitor → Update → Repeat
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 Data is a &lt;strong&gt;continuous process&lt;/strong&gt;, not a one-time step.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧪 Example (Simple but Real)
&lt;/h2&gt;

&lt;p&gt;Imagine a spam detection system:&lt;/p&gt;

&lt;h3&gt;
  
  
  Training data:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Clean emails
&lt;/li&gt;
&lt;li&gt;Proper grammar
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Production data:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Slang
&lt;/li&gt;
&lt;li&gt;Typos
&lt;/li&gt;
&lt;li&gt;Mixed languages
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Your model suddenly performs worse.&lt;/p&gt;

&lt;p&gt;Not because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The model is bad
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;The data changed&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  ⚙️ What You Should Do Instead
&lt;/h2&gt;

&lt;p&gt;Before choosing a model:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Audit your data
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;What do you actually have?
&lt;/li&gt;
&lt;li&gt;Is it usable?&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Simulate production inputs
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Test real-world scenarios
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Build data pipelines
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Collection
&lt;/li&gt;
&lt;li&gt;Cleaning
&lt;/li&gt;
&lt;li&gt;Transformation
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  4. Plan for updates
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;How will new data be added?&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧱 Data Pipelines Are the Real Foundation
&lt;/h2&gt;

&lt;p&gt;Your system should look like:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data Sources → Cleaning → Transformation → Storage → Model
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 If this pipeline is weak:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The system collapses
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;AI systems don’t improve because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You switch models
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They improve because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;You improve the data&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Don’t ask “Which model should we use?”&lt;br&gt;&lt;br&gt;
Ask: &lt;strong&gt;“Do we have the right data?”&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Anyone can download a model.&lt;/p&gt;

&lt;p&gt;Very few can:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Build and maintain high-quality data systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s where real advantage lies.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>Designing an AI System: Where Do You Even Start?</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Tue, 21 Apr 2026 07:30:00 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/designing-an-ai-system-where-do-you-even-start-5bfm</link>
      <guid>https://dev.to/siddhartha_reddy/designing-an-ai-system-where-do-you-even-start-5bfm</guid>
      <description>&lt;p&gt;"Most people start AI projects with models. That’s the wrong place to begin. Here’s how to think about designing AI systems correctly."&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Most AI projects fail before they even start.&lt;br&gt;&lt;br&gt;
Not because of bad models but because of bad starting points.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Default (Wrong) Approach
&lt;/h2&gt;

&lt;p&gt;Most people start like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Problem → Model → Data → Deployment
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ul&gt;
&lt;li&gt;Pick a model
&lt;/li&gt;
&lt;li&gt;Train it
&lt;/li&gt;
&lt;li&gt;Hope it works
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This leads to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Misaligned systems
&lt;/li&gt;
&lt;li&gt;Poor performance
&lt;/li&gt;
&lt;li&gt;Failed products
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧠 The Real Question
&lt;/h2&gt;

&lt;p&gt;Before you touch a model, ask:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;What problem are we actually solving?&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Not:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“Can we use AI here?”
&lt;/li&gt;
&lt;li&gt;“Which model should we use?”
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Who is the user?
&lt;/li&gt;
&lt;li&gt;What decision are we supporting?
&lt;/li&gt;
&lt;li&gt;What does success look like?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 AI is not the goal.&lt;br&gt;&lt;br&gt;
👉 The system solving a real problem is.&lt;/p&gt;


&lt;h2&gt;
  
  
  ⚙️ The Right Starting Point
&lt;/h2&gt;

&lt;p&gt;Real AI systems start like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Problem → Data → Constraints → System → Model
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Let’s break this down.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 1. Problem Definition
&lt;/h2&gt;

&lt;p&gt;Be specific.&lt;/p&gt;

&lt;p&gt;Bad:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“Build a recommendation system”&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Good:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“Increase user retention by recommending relevant content within 200ms”&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Constraints and goals matter more than the model.&lt;/p&gt;




&lt;h2&gt;
  
  
  📊 2. Understanding Data
&lt;/h2&gt;

&lt;p&gt;Ask:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What data do we have?
&lt;/li&gt;
&lt;li&gt;Is it reliable?
&lt;/li&gt;
&lt;li&gt;Does it reflect real-world usage?
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 No data → no system&lt;br&gt;&lt;br&gt;
👉 Bad data → bad system  &lt;/p&gt;


&lt;h2&gt;
  
  
  ⚠️ 3. Constraints (Most Ignored Step)
&lt;/h2&gt;

&lt;p&gt;Every system has constraints:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Latency (real-time vs batch)
&lt;/li&gt;
&lt;li&gt;Cost (compute limits)
&lt;/li&gt;
&lt;li&gt;Scale (number of users)
&lt;/li&gt;
&lt;li&gt;Accuracy requirements
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 These decisions shape EVERYTHING.&lt;/p&gt;


&lt;h2&gt;
  
  
  🧱 4. System Design
&lt;/h2&gt;

&lt;p&gt;Now you design:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data pipeline
&lt;/li&gt;
&lt;li&gt;Processing layers
&lt;/li&gt;
&lt;li&gt;Model integration
&lt;/li&gt;
&lt;li&gt;Serving layer
&lt;/li&gt;
&lt;li&gt;Monitoring
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This is where most real engineering happens.&lt;/p&gt;


&lt;h2&gt;
  
  
  🤖 5. Model Selection (Finally)
&lt;/h2&gt;

&lt;p&gt;Only now do you ask:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What model fits the constraints?
&lt;/li&gt;
&lt;li&gt;What tradeoffs are acceptable?
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Not:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“What’s the best model?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;“What’s the right model for this system?”&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;


&lt;h2&gt;
  
  
  🔁 Putting It Together
&lt;/h2&gt;

&lt;p&gt;The correct mindset:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Start with the problem  
Design the system  
Then choose the model
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;h2&gt;
  
  
  ⚠️ Why Most Teams Get This Wrong
&lt;/h2&gt;

&lt;p&gt;Because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Models are exciting
&lt;/li&gt;
&lt;li&gt;Systems are complex
&lt;/li&gt;
&lt;li&gt;Data is messy
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;So people skip to:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The easiest visible part, &lt;strong&gt;The Model&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;AI projects don’t fail because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Models are bad
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They fail because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;The system was never designed properly&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Don’t start with the model.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Start with the problem and constraints.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Anyone can train a model.&lt;/p&gt;

&lt;p&gt;Very few can:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Design a system that actually works&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s the difference.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>The Future: Engineers as AI System Architects</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Mon, 20 Apr 2026 07:30:00 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/the-future-engineers-as-ai-system-architects-4015</link>
      <guid>https://dev.to/siddhartha_reddy/the-future-engineers-as-ai-system-architects-4015</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;AI is not replacing engineers.&lt;br&gt;&lt;br&gt;
It’s redefining what it means to be one.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Narrative is Wrong
&lt;/h2&gt;

&lt;p&gt;The common belief:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“AI will replace programmers”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But what’s actually happening is more subtle:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AI is replacing parts of programming not the role of engineers&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 What’s Actually Changing
&lt;/h2&gt;

&lt;p&gt;Traditionally, engineers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Wrote code
&lt;/li&gt;
&lt;li&gt;Debugged logic
&lt;/li&gt;
&lt;li&gt;Built features
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now, increasingly, they:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Guide AI systems
&lt;/li&gt;
&lt;li&gt;Design workflows
&lt;/li&gt;
&lt;li&gt;Validate outputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The shift is not from &lt;em&gt;engineer → obsolete&lt;/em&gt;&lt;br&gt;&lt;br&gt;
👉 It’s from &lt;em&gt;engineer → architect&lt;/em&gt;&lt;/p&gt;


&lt;h2&gt;
  
  
  ⚙️ From Coding to Orchestration
&lt;/h2&gt;

&lt;p&gt;Old workflow:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Write → Debug → Deploy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;New workflow:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Define → Generate → Evaluate → Refine → Deploy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Engineers are no longer just writing code.&lt;/p&gt;

&lt;p&gt;They are:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Orchestrating systems that produce code&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🤖 The Rise of AI Systems
&lt;/h2&gt;

&lt;p&gt;Modern systems include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;LLMs
&lt;/li&gt;
&lt;li&gt;Tool integrations
&lt;/li&gt;
&lt;li&gt;Feedback loops
&lt;/li&gt;
&lt;li&gt;Memory layers
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These systems:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Generate solutions
&lt;/li&gt;
&lt;li&gt;Test them
&lt;/li&gt;
&lt;li&gt;Improve over time
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The intelligence is in the system, not just the model.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 The New Skill Stack
&lt;/h2&gt;

&lt;p&gt;The engineers who will thrive understand:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Systems thinking
&lt;/h3&gt;

&lt;p&gt;How components interact and fail&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Data flows
&lt;/h3&gt;

&lt;p&gt;Where inputs come from and how they change&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Evaluation
&lt;/h3&gt;

&lt;p&gt;How to measure correctness beyond accuracy&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Tooling
&lt;/h3&gt;

&lt;p&gt;How to integrate models into real workflows&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Feedback loops
&lt;/h3&gt;

&lt;p&gt;How systems improve over time&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ What Becomes Less Important
&lt;/h2&gt;

&lt;p&gt;Not irrelevant — but less dominant:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Writing boilerplate code
&lt;/li&gt;
&lt;li&gt;Memorizing syntax
&lt;/li&gt;
&lt;li&gt;Manual implementation of standard patterns
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 These are increasingly handled by AI.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧑‍💻 What Becomes More Important
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Designing robust systems
&lt;/li&gt;
&lt;li&gt;Handling uncertainty
&lt;/li&gt;
&lt;li&gt;Managing failure cases
&lt;/li&gt;
&lt;li&gt;Making trade-offs (speed vs cost vs accuracy)
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 These cannot be automated easily.&lt;/p&gt;




&lt;h2&gt;
  
  
  ⚠️ The Risk: Shallow Engineers
&lt;/h2&gt;

&lt;p&gt;There’s a danger emerging:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Engineers who can generate code…&lt;br&gt;&lt;br&gt;
but don’t understand systems&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;This leads to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fragile products
&lt;/li&gt;
&lt;li&gt;Poor debugging ability
&lt;/li&gt;
&lt;li&gt;Hidden technical debt
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 AI amplifies both skill and ignorance.&lt;/p&gt;




&lt;h2&gt;
  
  
  🚀 The Opportunity
&lt;/h2&gt;

&lt;p&gt;The engineers who win won’t be:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The best coders&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;They’ll be:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;The best system designers&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🔁 The Future of Software
&lt;/h2&gt;

&lt;p&gt;We’re moving toward systems that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Generate code
&lt;/li&gt;
&lt;li&gt;Test themselves
&lt;/li&gt;
&lt;li&gt;Adapt to new data
&lt;/li&gt;
&lt;li&gt;Improve continuously
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Software becomes:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;A living system, not a static artifact&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Learn how systems work.&lt;br&gt;&lt;br&gt;
Not just how to write code.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Final Thought
&lt;/h2&gt;

&lt;p&gt;Everyone is learning how to use AI.&lt;/p&gt;

&lt;p&gt;Very few are learning how to:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Design AI systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That gap defines the next generation of engineers.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>systemdesign</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>From Model to Product: Where AI Projects Actually Break</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Sun, 19 Apr 2026 07:30:00 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/from-model-to-product-where-ai-projects-actually-break-5dh8</link>
      <guid>https://dev.to/siddhartha_reddy/from-model-to-product-where-ai-projects-actually-break-5dh8</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;Most AI projects don’t fail in training.&lt;br&gt;&lt;br&gt;
They fail when you try to turn them into products.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Illusion: “The Model Works”
&lt;/h2&gt;

&lt;p&gt;You trained a model:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Good accuracy
&lt;/li&gt;
&lt;li&gt;Clean evaluation metrics
&lt;/li&gt;
&lt;li&gt;Solid results
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;So you think:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“We’re ready to ship.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But this is where most teams hit a wall.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧠 The Real Problem
&lt;/h2&gt;

&lt;p&gt;A working model ≠ a working product&lt;/p&gt;

&lt;p&gt;AI products require:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Reliability
&lt;/li&gt;
&lt;li&gt;Consistency
&lt;/li&gt;
&lt;li&gt;Usability
&lt;/li&gt;
&lt;li&gt;Trust
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 None of which are guaranteed by a model.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 1. The “Demo Trap”
&lt;/h2&gt;

&lt;p&gt;In demos:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Controlled inputs
&lt;/li&gt;
&lt;li&gt;Best-case scenarios
&lt;/li&gt;
&lt;li&gt;Clean outputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In production:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Messy inputs
&lt;/li&gt;
&lt;li&gt;Edge cases
&lt;/li&gt;
&lt;li&gt;Unpredictable behavior
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 What worked in a demo often breaks immediately in real usage.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 2. UX is an Afterthought
&lt;/h2&gt;

&lt;p&gt;Most AI systems are built like this:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Model first
&lt;/li&gt;
&lt;li&gt;UX later
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But users care about:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Response time
&lt;/li&gt;
&lt;li&gt;Clarity
&lt;/li&gt;
&lt;li&gt;Consistency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Not:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Your model architecture
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 A powerful model with poor UX feels broken.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 3. No Handling of Failure Cases
&lt;/h2&gt;

&lt;p&gt;AI systems WILL fail.&lt;/p&gt;

&lt;p&gt;But most products don’t plan for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Incorrect outputs
&lt;/li&gt;
&lt;li&gt;Uncertain predictions
&lt;/li&gt;
&lt;li&gt;Edge cases
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Good products:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Detect failure
&lt;/li&gt;
&lt;li&gt;Handle it gracefully
&lt;/li&gt;
&lt;li&gt;Communicate clearly
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This is product thinking, not model thinking.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 4. Latency Kills Experience
&lt;/h2&gt;

&lt;p&gt;Your model might be accurate…&lt;/p&gt;

&lt;p&gt;But if it takes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;2–3 seconds to respond
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Users feel:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“This is slow”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 Perception matters more than accuracy.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 5. Lack of Trust
&lt;/h2&gt;

&lt;p&gt;Users don’t trust AI by default.&lt;/p&gt;

&lt;p&gt;They need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Predictability
&lt;/li&gt;
&lt;li&gt;Transparency
&lt;/li&gt;
&lt;li&gt;Consistency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your system:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Sometimes works
&lt;/li&gt;
&lt;li&gt;Sometimes doesn’t
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Users stop relying on it.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 6. Integration is Harder Than Expected
&lt;/h2&gt;

&lt;p&gt;AI rarely exists alone.&lt;/p&gt;

&lt;p&gt;It must integrate with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Databases
&lt;/li&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;li&gt;Existing systems
&lt;/li&gt;
&lt;li&gt;Business workflows
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Most failures happen here, not in the model.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 7. Misaligned Expectations
&lt;/h2&gt;

&lt;p&gt;Stakeholders expect:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“Human-level intelligence”
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Probabilistic outputs
&lt;/li&gt;
&lt;li&gt;Imperfect predictions
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This gap kills projects.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 The Missing Layer
&lt;/h2&gt;

&lt;p&gt;Most teams focus on:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Model performance  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But ignore:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Product design&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧑‍💻 What Actually Works
&lt;/h2&gt;

&lt;p&gt;Successful AI products focus on:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ UX first
&lt;/h3&gt;

&lt;p&gt;Design around user experience&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Failure handling
&lt;/h3&gt;

&lt;p&gt;Expect and manage errors&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Speed optimization
&lt;/h3&gt;

&lt;p&gt;Balance latency vs accuracy&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Trust building
&lt;/h3&gt;

&lt;p&gt;Consistent behavior&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ System integration
&lt;/h3&gt;

&lt;p&gt;Fit into real workflows&lt;/p&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;A model answers:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Can this work?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;A product answers:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;“Will people actually use it?”&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;A great model doesn’t make a great product.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Great systems + UX do.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Everyone is building smarter models.&lt;/p&gt;

&lt;p&gt;Very few are building:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Better AI products&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s where the real impact is.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>product</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>Why 90% of ML Engineers Struggle in Real-World Systems</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Sat, 18 Apr 2026 07:30:00 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/why-90-of-ml-engineers-struggle-in-real-world-systems-19m</link>
      <guid>https://dev.to/siddhartha_reddy/why-90-of-ml-engineers-struggle-in-real-world-systems-19m</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;Most ML engineers don’t fail because they lack knowledge.&lt;br&gt;&lt;br&gt;
They fail because they’re solving the wrong problem.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Hard Truth
&lt;/h2&gt;

&lt;p&gt;Most ML engineers are trained to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Optimize models
&lt;/li&gt;
&lt;li&gt;Improve accuracy
&lt;/li&gt;
&lt;li&gt;Tune hyperparameters
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But real-world systems don’t fail because of bad models.&lt;/p&gt;

&lt;p&gt;They fail because of:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Bad system design&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 The Root Problem
&lt;/h2&gt;

&lt;p&gt;ML education focuses on:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Dataset → Model → Accuracy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;But real-world systems look like:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data → Pipeline → System → Monitoring → Feedback → Iteration
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 The model is just one part of a much bigger system&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 1. Too Much Focus on Accuracy
&lt;/h2&gt;

&lt;p&gt;Engineers obsess over:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;92% → 94% accuracy
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But ignore:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data quality
&lt;/li&gt;
&lt;li&gt;Pipeline reliability
&lt;/li&gt;
&lt;li&gt;System latency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 A slightly worse model in a solid system&lt;br&gt;&lt;br&gt;
will outperform a perfect model in a broken one.&lt;/p&gt;


&lt;h2&gt;
  
  
  ❌ 2. No Understanding of Data in Production
&lt;/h2&gt;

&lt;p&gt;In training:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean datasets
&lt;/li&gt;
&lt;li&gt;Well-structured inputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In production:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Missing values
&lt;/li&gt;
&lt;li&gt;Noisy inputs
&lt;/li&gt;
&lt;li&gt;Changing distributions
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Many engineers don’t design for this reality.&lt;/p&gt;


&lt;h2&gt;
  
  
  ❌ 3. Weak System Design Skills
&lt;/h2&gt;

&lt;p&gt;ML engineers often struggle with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;li&gt;Scalability
&lt;/li&gt;
&lt;li&gt;Distributed systems
&lt;/li&gt;
&lt;li&gt;Fault tolerance
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Because these aren’t taught in most ML paths.&lt;/p&gt;


&lt;h2&gt;
  
  
  ❌ 4. Ignoring the Pipeline
&lt;/h2&gt;

&lt;p&gt;They think:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“The model is the product”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But in reality:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;The pipeline is the product&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Problems appear in:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Preprocessing mismatch
&lt;/li&gt;
&lt;li&gt;Feature inconsistency
&lt;/li&gt;
&lt;li&gt;Data leakage
&lt;/li&gt;
&lt;/ul&gt;


&lt;h2&gt;
  
  
  ❌ 5. No Monitoring Mindset
&lt;/h2&gt;

&lt;p&gt;After deployment:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Train → Deploy → Done
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This is a mistake.&lt;/p&gt;

&lt;p&gt;Real systems require:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Monitor → Evaluate → Improve → Repeat
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 Without this, systems degrade silently.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 6. Poor Debugging Skills
&lt;/h2&gt;

&lt;p&gt;When models fail:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It’s not obvious why
&lt;/li&gt;
&lt;li&gt;It’s not reproducible
&lt;/li&gt;
&lt;li&gt;It’s not localized
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Debugging AI systems requires:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data tracing
&lt;/li&gt;
&lt;li&gt;Experiment tracking
&lt;/li&gt;
&lt;li&gt;System-level thinking
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This is very different from traditional debugging.&lt;/p&gt;




&lt;h2&gt;
  
  
  ❌ 7. No Product Thinking
&lt;/h2&gt;

&lt;p&gt;ML engineers often optimize for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Metrics
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But products require:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;User experience
&lt;/li&gt;
&lt;li&gt;Latency
&lt;/li&gt;
&lt;li&gt;Reliability
&lt;/li&gt;
&lt;li&gt;Business impact
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 A high-accuracy model that users don’t trust is useless.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 The Real Skill Gap
&lt;/h2&gt;

&lt;p&gt;It’s not:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“ML knowledge”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It’s:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Systems thinking&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧑‍💻 What Actually Makes a Strong ML Engineer
&lt;/h2&gt;

&lt;p&gt;The best engineers understand:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Data systems
&lt;/h3&gt;

&lt;p&gt;How data flows and breaks&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Pipelines
&lt;/h3&gt;

&lt;p&gt;End-to-end consistency&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Infrastructure
&lt;/h3&gt;

&lt;p&gt;Serving, scaling, latency&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Monitoring
&lt;/h3&gt;

&lt;p&gt;Real-world performance&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Feedback loops
&lt;/h3&gt;

&lt;p&gt;Continuous improvement&lt;/p&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;If you focus only on models:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You’ll stay stuck in notebooks&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;If you learn systems:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You’ll build real products&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;ML is not just about models.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;It’s about building reliable systems.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Most people are trying to become better at machine learning.&lt;/p&gt;

&lt;p&gt;Very few are trying to become:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Better at building AI systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s the difference.&lt;/p&gt;




</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>The Hidden Cost of AI Systems Nobody Talks About.</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Fri, 17 Apr 2026 07:30:00 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/the-hidden-cost-of-ai-systems-nobody-talks-about-2pgh</link>
      <guid>https://dev.to/siddhartha_reddy/the-hidden-cost-of-ai-systems-nobody-talks-about-2pgh</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;AI isn’t expensive.&lt;br&gt;&lt;br&gt;
Bad AI systems are.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💸 The Illusion: “AI is Cheap Now”
&lt;/h2&gt;

&lt;p&gt;With APIs and open-source models, it feels like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Spin up a model
&lt;/li&gt;
&lt;li&gt;Plug in an API
&lt;/li&gt;
&lt;li&gt;Ship a product
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Done.&lt;/p&gt;

&lt;p&gt;But that’s the &lt;strong&gt;demo illusion&lt;/strong&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  🚨 The Reality: Cost Starts After Deployment
&lt;/h2&gt;

&lt;p&gt;The real cost of AI systems doesn’t show up when you build them.&lt;/p&gt;

&lt;p&gt;It shows up when you:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Run them continuously in production&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  ⚙️ 1. Infrastructure Costs (The Silent Drain)
&lt;/h2&gt;

&lt;p&gt;Running AI at scale requires:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;GPUs / high-performance CPUs
&lt;/li&gt;
&lt;li&gt;Memory-heavy systems
&lt;/li&gt;
&lt;li&gt;Distributed infrastructure
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even simple systems:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Handle thousands of requests
&lt;/li&gt;
&lt;li&gt;Run models repeatedly
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Costs scale with &lt;strong&gt;usage, not development&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  ⏱️ 2. Latency vs Cost Tradeoff
&lt;/h2&gt;

&lt;p&gt;You want:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fast responses
&lt;/li&gt;
&lt;li&gt;High accuracy
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Faster models = more compute
&lt;/li&gt;
&lt;li&gt;Cheaper models = worse performance
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 You’re constantly balancing:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Speed ↔ Cost ↔ Accuracy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You can’t optimize all three.&lt;/p&gt;




&lt;h2&gt;
  
  
  🔁 3. Continuous Retraining
&lt;/h2&gt;

&lt;p&gt;Your model doesn’t stay good.&lt;/p&gt;

&lt;p&gt;You need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New data pipelines
&lt;/li&gt;
&lt;li&gt;Regular retraining
&lt;/li&gt;
&lt;li&gt;Validation cycles
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;More compute
&lt;/li&gt;
&lt;li&gt;More engineering time
&lt;/li&gt;
&lt;li&gt;More complexity
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 AI systems are &lt;strong&gt;never “done”&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🧑‍💻 4. Engineering Overhead
&lt;/h2&gt;

&lt;p&gt;The hidden cost isn’t just infra.&lt;/p&gt;

&lt;p&gt;It’s people.&lt;/p&gt;

&lt;p&gt;You need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;ML engineers
&lt;/li&gt;
&lt;li&gt;Data engineers
&lt;/li&gt;
&lt;li&gt;Backend engineers
&lt;/li&gt;
&lt;li&gt;DevOps / MLOps
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The model is 10% of the effort&lt;br&gt;&lt;br&gt;
👉 The system is 90%&lt;/p&gt;




&lt;h2&gt;
  
  
  🐛 5. Debugging is Expensive
&lt;/h2&gt;

&lt;p&gt;When AI systems fail:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It’s not obvious why
&lt;/li&gt;
&lt;li&gt;It’s not reproducible
&lt;/li&gt;
&lt;li&gt;It’s not localized
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Debugging requires:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Logs
&lt;/li&gt;
&lt;li&gt;Data tracing
&lt;/li&gt;
&lt;li&gt;Experiment tracking
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 This takes serious time.&lt;/p&gt;




&lt;h2&gt;
  
  
  📊 6. Monitoring &amp;amp; Observability
&lt;/h2&gt;

&lt;p&gt;To keep systems reliable, you need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Drift detection
&lt;/li&gt;
&lt;li&gt;Performance tracking
&lt;/li&gt;
&lt;li&gt;Alerting systems
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without this:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Your system degrades silently.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;With this:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You pay in infrastructure + engineering.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🔒 7. Risk &amp;amp; Reliability Costs
&lt;/h2&gt;

&lt;p&gt;AI introduces new risks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Incorrect predictions
&lt;/li&gt;
&lt;li&gt;Bias issues
&lt;/li&gt;
&lt;li&gt;Hallucinations
&lt;/li&gt;
&lt;li&gt;Edge-case failures
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To handle this, you need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Safeguards
&lt;/li&gt;
&lt;li&gt;Human-in-the-loop systems
&lt;/li&gt;
&lt;li&gt;Validation layers
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 More complexity = more cost&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 The Real Insight
&lt;/h2&gt;

&lt;p&gt;Most teams think:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“We need a better model”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But the real problem is:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;We need a better system&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;AI doesn’t become expensive because of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Model size
&lt;/li&gt;
&lt;li&gt;Training cost
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It becomes expensive because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;You have to run, maintain, and evolve the system&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;AI is not a one-time cost.&lt;br&gt;&lt;br&gt;
It’s a &lt;strong&gt;continuous system expense&lt;/strong&gt;.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Anyone can build an AI demo.&lt;/p&gt;

&lt;p&gt;Very few can afford to:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Run it reliably in production&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s the real challenge.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>Inside an AI Pipeline: What Actually Happens After You Train a Model</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Thu, 16 Apr 2026 20:45:28 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/inside-an-ai-pipeline-what-actually-happens-after-you-train-a-model-1gpj</link>
      <guid>https://dev.to/siddhartha_reddy/inside-an-ai-pipeline-what-actually-happens-after-you-train-a-model-1gpj</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;Training a model is the easiest part of AI.&lt;br&gt;&lt;br&gt;
Building the system around it is where things get real.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 The Biggest Misunderstanding in AI
&lt;/h2&gt;

&lt;p&gt;Most people think AI looks like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data → Model → Predictions
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That’s a &lt;strong&gt;toy version&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Real-world AI systems look like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data → Validation → Preprocessing → Feature Engineering → Model → Post-processing → Serving → Monitoring → Feedback → Retraining
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 The model is just &lt;strong&gt;one step in a long pipeline&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  ⚙️ Step 1: Data Ingestion
&lt;/h2&gt;

&lt;p&gt;Your system starts with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Databases
&lt;/li&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;li&gt;Logs
&lt;/li&gt;
&lt;li&gt;User input
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Problems:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Missing data
&lt;/li&gt;
&lt;li&gt;Inconsistent formats
&lt;/li&gt;
&lt;li&gt;Delayed updates
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 If your data is bad, everything downstream is broken.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧹 Step 2: Data Validation &amp;amp; Cleaning
&lt;/h2&gt;

&lt;p&gt;Before anything else:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Null checks
&lt;/li&gt;
&lt;li&gt;Schema validation
&lt;/li&gt;
&lt;li&gt;Outlier detection
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Age = -5
&lt;/li&gt;
&lt;li&gt;Salary = 999999999
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Garbage in → garbage out&lt;/p&gt;




&lt;h2&gt;
  
  
  🧪 Step 3: Preprocessing
&lt;/h2&gt;

&lt;p&gt;Transform raw data:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Normalization
&lt;/li&gt;
&lt;li&gt;Encoding
&lt;/li&gt;
&lt;li&gt;Tokenization
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;⚠️ Critical issue:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Training preprocessing ≠ Production preprocessing&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧩 Step 4: Feature Engineering
&lt;/h2&gt;

&lt;p&gt;This is where:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Domain knowledge meets ML&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Aggregations
&lt;/li&gt;
&lt;li&gt;Time-based features
&lt;/li&gt;
&lt;li&gt;Derived metrics
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🤖 Step 5: Model Training
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Train
&lt;/li&gt;
&lt;li&gt;Tune
&lt;/li&gt;
&lt;li&gt;Evaluate
&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;A great model inside a bad system still fails.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🔄 Step 6: Post-processing
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Thresholding
&lt;/li&gt;
&lt;li&gt;Ranking
&lt;/li&gt;
&lt;li&gt;Business rules
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🚀 Step 7: Model Serving
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;li&gt;Batch jobs
&lt;/li&gt;
&lt;li&gt;Streaming
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Challenges:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Latency
&lt;/li&gt;
&lt;li&gt;Scaling
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  📊 Step 8: Monitoring
&lt;/h2&gt;

&lt;p&gt;Track:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Accuracy
&lt;/li&gt;
&lt;li&gt;Input drift
&lt;/li&gt;
&lt;li&gt;Latency
&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Without monitoring, you’re flying blind.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  📉 Step 9: Feedback Loop
&lt;/h2&gt;

&lt;p&gt;Collect:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;User feedback
&lt;/li&gt;
&lt;li&gt;Errors
&lt;/li&gt;
&lt;li&gt;Edge cases
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Feed into retraining.&lt;/p&gt;




&lt;h2&gt;
  
  
  🔁 Step 10: Continuous Retraining
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;New Data → Retrain → Deploy → Repeat
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;h2&gt;
  
  
  🧩 Full Pipeline
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data Sources
     ↓
Validation
     ↓
Preprocessing
     ↓
Feature Engineering
     ↓
Model
     ↓
Post-processing
     ↓
Serving
     ↓
Monitoring
     ↓
Feedback
     ↓
Retraining
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;h2&gt;
  
  
  ⚠️ Where Systems Fail
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Data quality
&lt;/li&gt;
&lt;li&gt;Pipeline mismatch
&lt;/li&gt;
&lt;li&gt;No monitoring
&lt;/li&gt;
&lt;li&gt;No feedback
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;If you focus only on models:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You build demos  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;If you focus on pipelines:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You build products  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 Key Insight
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;The model is just a component.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;The pipeline is the product.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🔗 Series
&lt;/h2&gt;

&lt;p&gt;Previous:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;AI Doesn’t Write Code, Systems Do
&lt;/li&gt;
&lt;li&gt;Why Most AI Systems Fail in Production
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Next:&lt;br&gt;
👉 The Hidden Cost of AI Systems Nobody Talks About&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>"Why Most AI Systems Fail in Production (And No One Talks About It)"</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Thu, 16 Apr 2026 13:31:12 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/why-most-ai-systems-fail-in-production-and-no-one-talks-about-it-2i0b</link>
      <guid>https://dev.to/siddhartha_reddy/why-most-ai-systems-fail-in-production-and-no-one-talks-about-it-2i0b</guid>
      <description>&lt;h2&gt;
  
  
   "AI demos look perfect production systems don’t. Here’s why most AI systems fail in the real world."
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;AI demos look magical.&lt;br&gt;&lt;br&gt;
Production systems look broken.&lt;br&gt;&lt;br&gt;
And the gap between them is where most teams fail.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚨 The Truth Nobody Likes to Admit
&lt;/h2&gt;

&lt;p&gt;Most AI systems don’t fail in training.&lt;/p&gt;

&lt;p&gt;They fail in &lt;strong&gt;production&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Not because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The model is bad
&lt;/li&gt;
&lt;li&gt;The accuracy is low
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Real-world systems are messy, unpredictable, and constantly changing&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 The “Demo vs Reality” Problem
&lt;/h2&gt;

&lt;p&gt;In demos:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean datasets
&lt;/li&gt;
&lt;li&gt;Controlled inputs
&lt;/li&gt;
&lt;li&gt;No edge cases
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In production:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Noisy data
&lt;/li&gt;
&lt;li&gt;Missing values
&lt;/li&gt;
&lt;li&gt;Unexpected inputs
&lt;/li&gt;
&lt;li&gt;Changing distributions
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Your model isn’t solving the same problem anymore.&lt;/p&gt;




&lt;h2&gt;
  
  
  📉 1. Data Drift (Silent Killer)
&lt;/h2&gt;

&lt;p&gt;Your model was trained on &lt;strong&gt;past data&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Production gives you:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;New data, new patterns, new behavior&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Types of drift:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Feature drift (input changes)
&lt;/li&gt;
&lt;li&gt;Concept drift (relationship changes)
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fraud model trained on 2023 data
&lt;/li&gt;
&lt;li&gt;Used in 2025 → patterns completely different
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Accuracy drops silently.&lt;/p&gt;




&lt;h2&gt;
  
  
  ⚙️ 2. The Pipeline is the Real System
&lt;/h2&gt;

&lt;p&gt;Most people focus on the model.&lt;/p&gt;

&lt;p&gt;But the real system is:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Data → Preprocessing → Model → Post-processing → API → Monitoring
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Failure can happen anywhere:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Wrong preprocessing
&lt;/li&gt;
&lt;li&gt;Feature mismatch
&lt;/li&gt;
&lt;li&gt;Data leakage
&lt;/li&gt;
&lt;li&gt;Version mismatch
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The model is just one piece.&lt;/p&gt;




&lt;h2&gt;
  
  
  🐛 3. Edge Cases Destroy Everything
&lt;/h2&gt;

&lt;p&gt;AI works well on:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Common cases”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But production is full of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Rare inputs
&lt;/li&gt;
&lt;li&gt;Unexpected formats
&lt;/li&gt;
&lt;li&gt;Adversarial cases
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;NLP model trained on clean text
&lt;/li&gt;
&lt;li&gt;Production input = slang + emojis + typos
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 System breaks instantly.&lt;/p&gt;




&lt;h2&gt;
  
  
  ⏱️ 4. Latency &amp;amp; Cost Constraints
&lt;/h2&gt;

&lt;p&gt;Your model works great…&lt;/p&gt;

&lt;p&gt;Until:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It takes 2 seconds per request
&lt;/li&gt;
&lt;li&gt;Or costs too much to run
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Production requires:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Low latency
&lt;/li&gt;
&lt;li&gt;High throughput
&lt;/li&gt;
&lt;li&gt;Cost efficiency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 A perfect model that’s slow is useless.&lt;/p&gt;




&lt;h2&gt;
  
  
  🔁 5. No Feedback Loop = Slow Death
&lt;/h2&gt;

&lt;p&gt;Most systems are deployed like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Train → Deploy → Forget
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That’s a mistake.&lt;/p&gt;

&lt;p&gt;Real systems need:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Monitor → Evaluate → Retrain → Improve
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Without feedback:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Performance degrades
&lt;/li&gt;
&lt;li&gt;Errors accumulate
&lt;/li&gt;
&lt;li&gt;Users lose trust
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧩 6. Observability is Missing
&lt;/h2&gt;

&lt;p&gt;Most teams don’t track:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Model performance in real-time
&lt;/li&gt;
&lt;li&gt;Input distributions
&lt;/li&gt;
&lt;li&gt;Failure cases
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;So when things break:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You don’t even know why.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🤖 The Real Problem
&lt;/h2&gt;

&lt;p&gt;The biggest mistake teams make:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Treating AI as a model problem&lt;br&gt;&lt;br&gt;
Instead of a &lt;strong&gt;systems problem&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧑‍💻 What Actually Works
&lt;/h2&gt;

&lt;p&gt;Successful AI systems focus on:&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Data pipelines
&lt;/h3&gt;

&lt;p&gt;Clean, versioned, monitored&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Continuous evaluation
&lt;/h3&gt;

&lt;p&gt;Not just offline metrics&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ Feedback loops
&lt;/h3&gt;

&lt;p&gt;Real-world learning&lt;/p&gt;

&lt;h3&gt;
  
  
  ✅ System design
&lt;/h3&gt;

&lt;p&gt;Not just model tuning&lt;/p&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;AI doesn’t fail because models are bad.&lt;/p&gt;

&lt;p&gt;It fails because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Systems are incomplete&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Building the model is easy.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Building the system is the real challenge.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Everyone is building AI models.&lt;/p&gt;

&lt;p&gt;Very few are building:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Reliable AI systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 That’s where the real opportunity is.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>mlops</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>AI Doesn’t Write Code, Systems Do (And Most People Are Missing This)</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Tue, 14 Apr 2026 19:03:23 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/ai-doesnt-write-code-systems-do-and-most-people-are-missing-this-35hb</link>
      <guid>https://dev.to/siddhartha_reddy/ai-doesnt-write-code-systems-do-and-most-people-are-missing-this-35hb</guid>
      <description>&lt;h2&gt;
  
  
  "AI isn’t writing your code, systems are. Here’s what most developers are missing about AI coding."
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Everyone says AI is replacing programmers.&lt;br&gt;&lt;br&gt;
That’s not what’s happening.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Something much more interesting is.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  ⚠️ The Biggest Misconception in AI Right Now
&lt;/h2&gt;

&lt;p&gt;People think:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“AI writes code.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;What’s actually happening:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Systems built around AI are writing software.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And if you don’t understand that difference, you’re going to fall behind fast.&lt;/p&gt;




&lt;h2&gt;
  
  
  🧠 The Illusion of Intelligence
&lt;/h2&gt;

&lt;p&gt;When you use AI to generate code, it &lt;em&gt;feels&lt;/em&gt; like you're working with something intelligent.&lt;/p&gt;

&lt;p&gt;But under the hood, it’s just:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Predicting the next token
&lt;/li&gt;
&lt;li&gt;Based on patterns
&lt;/li&gt;
&lt;li&gt;Without understanding correctness
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It doesn’t “know” your code works.&lt;/p&gt;

&lt;p&gt;It only knows:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“This looks like code that usually follows this prompt.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;So why does it work so well?&lt;/p&gt;




&lt;h2&gt;
  
  
  ⚙️ Why AI Code Actually Works (Surprisingly Well)
&lt;/h2&gt;

&lt;p&gt;AI works in coding not because it's smart&lt;br&gt;&lt;br&gt;
but because &lt;strong&gt;software engineering is structured for it to succeed&lt;/strong&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  1. Code is predictable
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Repeated patterns
&lt;/li&gt;
&lt;li&gt;Standard libraries
&lt;/li&gt;
&lt;li&gt;Known structures
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;
  
  
  2. Feedback loops are instant
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Compile → fail → fix
&lt;/li&gt;
&lt;li&gt;Test → fail → fix
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;
  
  
  3. “Good enough” wins
&lt;/h3&gt;

&lt;p&gt;Companies don’t need perfect code. They need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Faster shipping
&lt;/li&gt;
&lt;li&gt;Lower costs
&lt;/li&gt;
&lt;li&gt;Acceptable reliability
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 AI fits perfectly into this system.&lt;/p&gt;


&lt;h2&gt;
  
  
  💥 Where AI Completely Breaks
&lt;/h2&gt;

&lt;p&gt;AI is great at &lt;strong&gt;local problems&lt;/strong&gt;, but fails at &lt;strong&gt;system-level thinking&lt;/strong&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  ❌ Long-term architecture
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Doesn’t plan systems
&lt;/li&gt;
&lt;li&gt;Doesn’t maintain consistency
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;
  
  
  ❌ State &amp;amp; memory
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;No real awareness of past decisions
&lt;/li&gt;
&lt;li&gt;No persistent understanding
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;
  
  
  ❌ Debugging complex systems
&lt;/h3&gt;

&lt;p&gt;AI can fix:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Syntax errors
&lt;/li&gt;
&lt;li&gt;Small bugs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But fails at:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Distributed failures
&lt;/li&gt;
&lt;li&gt;Race conditions
&lt;/li&gt;
&lt;li&gt;Deep system issues
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Because these require &lt;strong&gt;causal reasoning&lt;/strong&gt;, not pattern matching.&lt;/p&gt;


&lt;h2&gt;
  
  
  🧩 The Real Architecture of “AI Coding”
&lt;/h2&gt;

&lt;p&gt;AI coding tools are NOT just models.&lt;/p&gt;

&lt;p&gt;They are &lt;strong&gt;systems&lt;/strong&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  🧠 What’s actually happening:
&lt;/h3&gt;


&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;User Prompt
     ↓
LLM (generates code)
     ↓
Tooling Layer
 (compiler / tests / linters)
     ↓
Feedback Loop
 (errors, logs, outputs)
     ↓
Iteration Engine
 (fix → retry → improve)
     ↓
Final Output
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;👉 The intelligence is NOT in the model&lt;br&gt;&lt;br&gt;
👉 The intelligence is in the &lt;strong&gt;loop&lt;/strong&gt;&lt;/p&gt;


&lt;h2&gt;
  
  
  🤖 Agentic Systems: The Real Shift
&lt;/h2&gt;

&lt;p&gt;We’re moving from:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Prompt → Output
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;To:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Goal → Plan → Execute → Evaluate → Iterate
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This is &lt;strong&gt;agentic coding&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;These systems:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Write code
&lt;/li&gt;
&lt;li&gt;Run it
&lt;/li&gt;
&lt;li&gt;Analyze failures
&lt;/li&gt;
&lt;li&gt;Fix it
&lt;/li&gt;
&lt;li&gt;Repeat
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Until it works.&lt;/p&gt;

&lt;p&gt;👉 The model is just a component&lt;br&gt;&lt;br&gt;
👉 The &lt;strong&gt;system does the thinking&lt;/strong&gt;&lt;/p&gt;


&lt;h2&gt;
  
  
  🧑‍💻 What Happens to Engineers?
&lt;/h2&gt;

&lt;p&gt;You’re not being replaced.&lt;/p&gt;

&lt;p&gt;Your role is being &lt;strong&gt;redefined&lt;/strong&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  Old role:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Write code
&lt;/li&gt;
&lt;li&gt;Debug manually
&lt;/li&gt;
&lt;li&gt;Build features
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;
  
  
  New role:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Design systems
&lt;/li&gt;
&lt;li&gt;Orchestrate AI workflows
&lt;/li&gt;
&lt;li&gt;Validate outputs
&lt;/li&gt;
&lt;li&gt;Own complexity
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 The best engineers won’t be:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“The fastest coders”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;👉 They’ll be:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;The best system designers&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;


&lt;h2&gt;
  
  
  ⚠️ The Hidden Problem Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;AI introduces a dangerous shift:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You didn’t write the code…&lt;br&gt;&lt;br&gt;
&lt;strong&gt;But you’re still responsible for it.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;This leads to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Shallow understanding
&lt;/li&gt;
&lt;li&gt;Fragile systems
&lt;/li&gt;
&lt;li&gt;Hidden technical debt
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you rely blindly on AI, you lose:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Code intuition&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And that’s where things break.&lt;/p&gt;


&lt;h2&gt;
  
  
  🔁 The Future: Software as a Feedback Loop
&lt;/h2&gt;

&lt;p&gt;We’re moving toward:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Generate → Test → Fix → Deploy → Monitor → Repeat
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Continuously.&lt;/p&gt;

&lt;p&gt;Software won’t be written once.&lt;/p&gt;

&lt;p&gt;It will be:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Continuously generated and refined by systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🚀 Final Take
&lt;/h2&gt;

&lt;p&gt;The narrative is wrong.&lt;/p&gt;

&lt;p&gt;It’s not:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“AI writes code”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It’s:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;AI generates possibilities&lt;br&gt;&lt;br&gt;
Systems validate them&lt;br&gt;&lt;br&gt;
Engineers make them meaningful  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  🧠 If You Take One Thing Away
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Don’t focus on prompts.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Focus on systems.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;That’s where the real leverage is.&lt;/p&gt;




&lt;h2&gt;
  
  
  💬 Closing Thought
&lt;/h2&gt;

&lt;p&gt;Most people are learning how to &lt;em&gt;use&lt;/em&gt; AI.&lt;/p&gt;

&lt;p&gt;Very few are learning how AI systems actually &lt;em&gt;work&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;👉 That gap is your opportunity.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>systemdesign</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>When Can You Actually Trust a Machine Learning Model?</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Wed, 01 Apr 2026 12:29:13 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/when-can-you-actually-trust-a-machine-learning-model-27kh</link>
      <guid>https://dev.to/siddhartha_reddy/when-can-you-actually-trust-a-machine-learning-model-27kh</guid>
      <description>&lt;p&gt;Building a machine learning model is relatively straightforward today.&lt;/p&gt;

&lt;p&gt;You train it.&lt;br&gt;
Evaluate it.&lt;br&gt;
Tune it.&lt;/p&gt;

&lt;p&gt;Eventually, you get a model that performs well.&lt;br&gt;
But a more difficult question comes after:&lt;br&gt;
&lt;code&gt;Can you trust it?&lt;/code&gt;&lt;br&gt;
Not occasionally.&lt;br&gt;
Not in controlled environments.&lt;br&gt;
But consistently in the real world.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Illusion of Trust
&lt;/h2&gt;

&lt;p&gt;Many people assume trust comes from metrics.&lt;br&gt;
If a model has:&lt;br&gt;
&lt;code&gt;Accuracy: 94%&lt;/code&gt;&lt;br&gt;
It feels reliable.&lt;br&gt;
But accuracy doesn’t tell you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;when the model will fail&lt;/li&gt;
&lt;li&gt;how it will fail&lt;/li&gt;
&lt;li&gt;how often it fails in critical cases&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A model can be highly accurate and still be unreliable.&lt;br&gt;
Trust is not a number.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Trust Actually Means
&lt;/h2&gt;

&lt;p&gt;In machine learning, trust is not about perfection.&lt;br&gt;
It’s about &lt;strong&gt;predictability&lt;/strong&gt;.&lt;br&gt;
A trustworthy model is one that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;behaves consistently&lt;/li&gt;
&lt;li&gt;fails in expected ways&lt;/li&gt;
&lt;li&gt;performs reliably across conditions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It doesn’t need to be perfect.&lt;br&gt;
It needs to be &lt;strong&gt;understandable in its behavior.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  When You Should Not Trust a Model
&lt;/h2&gt;

&lt;p&gt;There are clear situations where trust breaks down.&lt;br&gt;
&lt;strong&gt;1. When the data changes&lt;/strong&gt;&lt;br&gt;
If the model sees data that is different from training data:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;new patterns&lt;/li&gt;
&lt;li&gt;new distributions&lt;/li&gt;
&lt;li&gt;new environments&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All guarantees disappear.&lt;br&gt;
The model is now operating outside its experience.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. When edge cases matter&lt;/strong&gt;&lt;br&gt;
Models are optimized for average performance.&lt;br&gt;
They are not optimized for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;rare events&lt;/li&gt;
&lt;li&gt;unusual inputs&lt;/li&gt;
&lt;li&gt;extreme scenarios&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your system depends on edge-case correctness, trust becomes fragile.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. When the cost of failure is high&lt;/strong&gt;&lt;br&gt;
In some applications:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;healthcare&lt;/li&gt;
&lt;li&gt;finance&lt;/li&gt;
&lt;li&gt;safety systems&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even small errors are unacceptable.&lt;br&gt;
In these cases, trust must be extremely high — and rarely comes from the model alone.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. When the model is a black box&lt;/strong&gt;&lt;br&gt;
If you cannot understand:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;why predictions are made&lt;/li&gt;
&lt;li&gt;what features matter&lt;/li&gt;
&lt;li&gt;how decisions change&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then trust is limited.&lt;br&gt;
Opacity reduces confidence.&lt;/p&gt;

&lt;h2&gt;
  
  
  Signals of a Trustworthy Model
&lt;/h2&gt;

&lt;p&gt;Trust doesn’t come from a single metric.&lt;br&gt;
It comes from multiple signals.&lt;/p&gt;

&lt;h2&gt;
  
  
  Consistency across datasets
&lt;/h2&gt;

&lt;p&gt;The model performs similarly on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;training data&lt;/li&gt;
&lt;li&gt;validation data&lt;/li&gt;
&lt;li&gt;new real-world data&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Large gaps are a warning sign.&lt;/p&gt;

&lt;h2&gt;
  
  
  Stability under small changes
&lt;/h2&gt;

&lt;p&gt;If small input changes cause large output changes, the model is fragile.&lt;br&gt;
Stable models behave predictably under minor variations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Clear failure patterns
&lt;/h2&gt;

&lt;p&gt;You should be able to say:&lt;br&gt;
&lt;code&gt;“The model struggles in these specific situations.”&lt;/code&gt;&lt;br&gt;
If failures feel random, trust is low.&lt;/p&gt;

&lt;h2&gt;
  
  
  Continuous monitoring
&lt;/h2&gt;

&lt;p&gt;Trust is not static.&lt;br&gt;
Models degrade over time.&lt;br&gt;
A trustworthy system includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;monitoring&lt;/li&gt;
&lt;li&gt;alerts&lt;/li&gt;
&lt;li&gt;retraining strategies&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The System Around the Model Matters More
&lt;/h2&gt;

&lt;p&gt;A key insight:&lt;br&gt;
&lt;code&gt;Trust is not a property of the model. It’s a property of the system around it.&lt;/code&gt;&lt;br&gt;
A reliable ML system includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validation pipelines&lt;/li&gt;
&lt;li&gt;fallback mechanisms&lt;/li&gt;
&lt;li&gt;human oversight (when needed)&lt;/li&gt;
&lt;li&gt;monitoring and retraining&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even a strong model without these is risky.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Mental Shift
&lt;/h2&gt;

&lt;p&gt;Instead of asking:&lt;br&gt;
&lt;code&gt;“Is this model accurate?”&lt;/code&gt;&lt;br&gt;
Ask:&lt;br&gt;
&lt;code&gt;“When will this model fail, and how bad will that be?”&lt;/code&gt;&lt;br&gt;
This question leads to better decisions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;Machine learning models are powerful.&lt;br&gt;
But they are not inherently trustworthy.&lt;br&gt;
Trust is built through:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;understanding behavior&lt;/li&gt;
&lt;li&gt;testing limits&lt;/li&gt;
&lt;li&gt;designing systems around failure
The goal is not to build models that never fail.
The goal is to build systems where failure is &lt;strong&gt;expected, understood, and controlled.&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>ai</category>
      <category>python</category>
    </item>
    <item>
      <title>Why Your Machine Learning Model Breaks When Nothing Seems Wrong?</title>
      <dc:creator>Siddhartha Reddy</dc:creator>
      <pubDate>Tue, 31 Mar 2026 13:16:58 +0000</pubDate>
      <link>https://dev.to/siddhartha_reddy/why-your-machine-learning-model-breaks-when-nothing-seems-wrong-19o7</link>
      <guid>https://dev.to/siddhartha_reddy/why-your-machine-learning-model-breaks-when-nothing-seems-wrong-19o7</guid>
      <description>&lt;p&gt;You trained your model.&lt;/p&gt;

&lt;p&gt;The accuracy looked good.&lt;br&gt;
Validation results were consistent.&lt;br&gt;
The pipeline ran without errors.&lt;/p&gt;

&lt;p&gt;Everything suggested the model was ready.&lt;/p&gt;

&lt;p&gt;Then you used it in a real scenario.&lt;/p&gt;

&lt;p&gt;And it started failing.&lt;/p&gt;

&lt;p&gt;Not catastrophically.&lt;br&gt;
Not obviously.&lt;/p&gt;

&lt;p&gt;Just… wrong in ways that didn’t make sense.&lt;/p&gt;

&lt;p&gt;The confusing part?&lt;br&gt;
&lt;code&gt;Nothing in your code changed.&lt;/code&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Hidden Assumption Behind Every Model
&lt;/h2&gt;

&lt;p&gt;Every machine learning model relies on a quiet assumption:&lt;br&gt;
&lt;code&gt;The data in the future will look like the data in the past&lt;/code&gt;&lt;br&gt;
This assumption is rarely stated.&lt;/p&gt;

&lt;p&gt;But everything depends on it.&lt;/p&gt;

&lt;p&gt;When it holds, models perform well.&lt;br&gt;
When it breaks, models fail even if everything else is correct.&lt;/p&gt;

&lt;h2&gt;
  
  
  When Reality Doesn’t Match Training
&lt;/h2&gt;

&lt;p&gt;In practice, data is never static.&lt;br&gt;
It changes over time:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;user behavior evolves&lt;/li&gt;
&lt;li&gt;environments shift&lt;/li&gt;
&lt;li&gt;input formats vary&lt;/li&gt;
&lt;li&gt;noise increases&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is known as distribution shift.&lt;br&gt;
The model was trained on one distribution of data.&lt;br&gt;
It is now being used on another.&lt;br&gt;
The model hasn’t changed.&lt;br&gt;
But the world around it has.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Failure Is Hard to Detect
&lt;/h2&gt;

&lt;p&gt;Unlike code errors, this kind of failure is silent.&lt;br&gt;
There is no exception.&lt;br&gt;
No crash.&lt;br&gt;
No warning.&lt;br&gt;
The model continues to produce outputs.&lt;br&gt;
They just become:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;less accurate&lt;/li&gt;
&lt;li&gt;less consistent&lt;/li&gt;
&lt;li&gt;less reliable&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Because the model still “works,” the issue often goes unnoticed until it becomes serious.&lt;/p&gt;

&lt;h2&gt;
  
  
  Small Changes, Big Impact
&lt;/h2&gt;

&lt;p&gt;The most dangerous shifts are subtle.&lt;br&gt;
Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;slightly different lighting in images&lt;/li&gt;
&lt;li&gt;new categories of input data&lt;/li&gt;
&lt;li&gt;changes in user input patterns&lt;/li&gt;
&lt;li&gt;minor formatting differences
To a human, these changes seem trivial.
To a model, they can completely alter predictions.
Because models depend on patterns, even small changes can break those patterns.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Illusion of Stability
&lt;/h2&gt;

&lt;p&gt;During training and validation, everything looks stable.&lt;br&gt;
That’s because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;training data is consistent&lt;/li&gt;
&lt;li&gt;validation data comes from the same distribution&lt;/li&gt;
&lt;li&gt;assumptions are preserved&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The model is tested in an environment that mirrors its training conditions.&lt;br&gt;
But real-world data rarely behaves that way.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why More Accuracy Doesn’t Fix This
&lt;/h2&gt;

&lt;p&gt;Improving accuracy does not solve this problem.&lt;br&gt;
You can have:&lt;br&gt;
&lt;code&gt;95% validation accuracy&lt;/code&gt;&lt;br&gt;
And still fail in production.&lt;br&gt;
Because accuracy measures performance &lt;strong&gt;within a fixed dataset&lt;/strong&gt;.&lt;br&gt;
It does not measure:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;robustness&lt;/li&gt;
&lt;li&gt;adaptability&lt;/li&gt;
&lt;li&gt;resilience to change&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Real Problem: Static Models in a Dynamic World
&lt;/h2&gt;

&lt;p&gt;Machine learning models are static after training.&lt;br&gt;
The world is not.&lt;br&gt;
This mismatch creates failure.&lt;br&gt;
The model cannot adapt unless:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;it is retrained&lt;/li&gt;
&lt;li&gt;it is updated&lt;/li&gt;
&lt;li&gt;it is monitored&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without this, performance naturally degrades over time.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Recognize This Early
&lt;/h2&gt;

&lt;p&gt;Some warning signs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;performance slowly declines&lt;/li&gt;
&lt;li&gt;edge cases increase&lt;/li&gt;
&lt;li&gt;predictions become inconsistent&lt;/li&gt;
&lt;li&gt;certain inputs fail repeatedly
If the model worked before and now behaves differently, the issue may not be the model.
It may be the data distribution.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What Helps (But Doesn’t Eliminate the Problem)
&lt;/h2&gt;

&lt;p&gt;To reduce this risk:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;monitor model performance over time&lt;/li&gt;
&lt;li&gt;evaluate on fresh, real-world data&lt;/li&gt;
&lt;li&gt;retrain periodically&lt;/li&gt;
&lt;li&gt;design validation sets carefully&lt;/li&gt;
&lt;li&gt;test on slightly different distributions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These don’t eliminate the problem.&lt;br&gt;
But they make it visible.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Mental Shift
&lt;/h2&gt;

&lt;p&gt;Most people think:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;“If the model is good, it will keep working.”&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;A more accurate view is:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;A model is only as good as the data it was trained on — and how similar future data is to it.&lt;/code&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;Machine learning models don’t usually fail because something broke.&lt;br&gt;
They fail because something changed.&lt;br&gt;
And often, that change is subtle enough to go unnoticed until the model is no longer reliable.&lt;/p&gt;

&lt;p&gt;**Understanding this is the difference between building models that work once…&lt;/p&gt;

&lt;p&gt;…and systems that keep working over time.&lt;br&gt;
**&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>ai</category>
      <category>discuss</category>
    </item>
  </channel>
</rss>
