<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ashish Nadar</title>
    <description>The latest articles on DEV Community by Ashish Nadar (@ashish_nadar).</description>
    <link>https://dev.to/ashish_nadar</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3834591%2F192e420a-23da-40e1-ad7a-7061dd92dec0.png</url>
      <title>DEV Community: Ashish Nadar</title>
      <link>https://dev.to/ashish_nadar</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ashish_nadar"/>
    <language>en</language>
    <item>
      <title>Closing the Gap Between SCA Tools and Runtime Reality — Ashish Nadar</title>
      <dc:creator>Ashish Nadar</dc:creator>
      <pubDate>Fri, 10 Apr 2026 09:30:00 +0000</pubDate>
      <link>https://dev.to/ashish_nadar/closing-the-gap-between-sca-tools-and-runtime-reality-ashish-nadar-41g7</link>
      <guid>https://dev.to/ashish_nadar/closing-the-gap-between-sca-tools-and-runtime-reality-ashish-nadar-41g7</guid>
      <description>&lt;p&gt;The alert came in on a Tuesday morning.&lt;/p&gt;

&lt;p&gt;A critical CVE. Severity score 9.8. Affecting one of the most widely used&lt;br&gt;
open-source libraries in the Node.js ecosystem.&lt;/p&gt;

&lt;p&gt;Our team had Snyk. We had Wiz. We had automated scanning pipelines and&lt;br&gt;
weekly vulnerability reports. By most measures, we were well-equipped.&lt;/p&gt;

&lt;p&gt;So when the question landed in the security channel —&lt;br&gt;
&lt;em&gt;"Where are we actually exposed in production right now?"&lt;/em&gt; —&lt;br&gt;
we assumed the answer would take minutes.&lt;/p&gt;

&lt;p&gt;It took most of the day.&lt;/p&gt;

&lt;p&gt;That gap — between the tooling we had and the confidence we needed —&lt;br&gt;
is exactly what this article is about.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Core Problem
&lt;/h2&gt;

&lt;p&gt;SCA tools like Snyk and Wiz are excellent at what they do.&lt;br&gt;
They continuously scan repositories and flag vulnerable dependencies.&lt;/p&gt;

&lt;p&gt;But they scan &lt;strong&gt;source code&lt;/strong&gt;. Not production.&lt;/p&gt;

&lt;p&gt;And in any sufficiently complex environment, those two things&lt;br&gt;
can look very different:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Deployments routinely lag behind the latest commit&lt;/li&gt;
&lt;li&gt;Dev dependencies appear in source but never reach runtime&lt;/li&gt;
&lt;li&gt;Inactive repos still show up in scans — inflating apparent exposure&lt;/li&gt;
&lt;li&gt;Different environments run different versions of the same service&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The result? When a critical CVE drops, your SCA tool gives you&lt;br&gt;
a list of repositories where the vulnerable library &lt;em&gt;might&lt;/em&gt; exist.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What it cannot tell you is whether that library is actually deployed.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The Fix: A Runtime-First Approach
&lt;/h2&gt;

&lt;p&gt;The most reliable source of truth isn't your source repository.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;It's the code actually running in production.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;For AWS Lambda, this is surprisingly practical. Lambda functions&lt;br&gt;
contain their packaged dependencies — the actual &lt;code&gt;node_modules&lt;/code&gt;&lt;br&gt;
bundled at deployment time. Inspecting these directly gives you&lt;br&gt;
immediate, high-confidence answers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which functions are affected?&lt;/li&gt;
&lt;li&gt;Which version is running?&lt;/li&gt;
&lt;li&gt;Which environment?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No cross-team coordination. No waiting. No uncertainty.&lt;/p&gt;




&lt;h2&gt;
  
  
  What's in the Full Article
&lt;/h2&gt;

&lt;p&gt;The complete research article covers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;📖 The real incident that revealed this gap&lt;/li&gt;
&lt;li&gt;🔍 Why team-based verification fails under pressure&lt;/li&gt;
&lt;li&gt;🧪 A practical runtime inspection workflow for AWS Lambda&lt;/li&gt;
&lt;li&gt;🔄 How this pattern extends to containers and EC2&lt;/li&gt;
&lt;li&gt;🏗️ Building this as a permanent MCP-powered capability&lt;/li&gt;
&lt;li&gt;📊 How SCA and runtime inspection work together&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Read the Full Article
&lt;/h2&gt;

&lt;p&gt;This is a glimpse. The complete article — with detailed workflows,&lt;br&gt;
diagrams, and implementation patterns — is published on Medium.&lt;/p&gt;

&lt;h3&gt;
  
  
  👉 &lt;a href="https://medium.com/@ashish-nadar/ashish-nadar-scholar-sca-tools-runtime-reality-f6c094a62619" rel="noopener noreferrer"&gt;Read the full article here&lt;/a&gt;
&lt;/h3&gt;

</description>
      <category>security</category>
      <category>devops</category>
      <category>cloud</category>
      <category>programming</category>
    </item>
    <item>
      <title>The Day a Green Pipeline Told Me Everything Was Fine — And It Wasn't</title>
      <dc:creator>Ashish Nadar</dc:creator>
      <pubDate>Fri, 20 Mar 2026 08:30:00 +0000</pubDate>
      <link>https://dev.to/ashish_nadar/the-day-a-green-pipeline-told-me-everything-was-fine-and-it-wasnt-336</link>
      <guid>https://dev.to/ashish_nadar/the-day-a-green-pipeline-told-me-everything-was-fine-and-it-wasnt-336</guid>
      <description>&lt;p&gt;&lt;em&gt;By Ashish Nadar — Scholar in Secure Cloud · Operational Excellence · AI-Enabled Software&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;There's a moment every researcher dreads — not the moment when something breaks loudly and obviously, but the moment when everything appears to be working perfectly, and you have a quiet, creeping feeling that it isn't.&lt;/p&gt;

&lt;p&gt;I had that moment midway through one of my research projects on AI-assisted backend development.&lt;/p&gt;

&lt;p&gt;The pipeline was green. The tests were passing. The code was clean and well-structured. By every standard metric, the system looked healthy.&lt;/p&gt;

&lt;p&gt;But something had shifted. Silently.&lt;/p&gt;

&lt;p&gt;Underneath all those green checkmarks, a fallback path that should have triggered under a specific failure condition had quietly stopped working. Nobody was careless. The AI-generated code was genuinely good. And yet the system was misbehaving in a way none of our internal tests had caught.&lt;/p&gt;

&lt;p&gt;That incident led me to one of the most important questions I've explored in my research:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;In a world where AI can build and test a system in the same breath — what does independent validation even mean?&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  The Problem: Behavioral Drift
&lt;/h2&gt;

&lt;p&gt;When AI generates your backend logic &lt;strong&gt;and&lt;/strong&gt; updates your tests in the same workflow, something subtle and dangerous happens.&lt;/p&gt;

&lt;p&gt;The system changes. The tests are updated to validate the new behavior. The pipeline reports green. And nobody notices that the "new behavior" being validated is actually &lt;strong&gt;incorrect&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;I call this &lt;strong&gt;behavioral drift&lt;/strong&gt; — and it's one of the most invisible failure modes in modern AI-assisted development.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1ziyzlax81l3m5xq5185.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1ziyzlax81l3m5xq5185.png" alt="When AI modifies logic and tests in the same pass, the pipeline stays green — but the system quietly gets it wrong." width="800" height="660"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The scariest part? The code looks cleaner after the AI refactor. Better structured. More readable. And yet the system is quietly doing the wrong thing.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Fix: An Independent Validation Layer
&lt;/h2&gt;

&lt;p&gt;The solution isn't better internal tests. It's putting the tests &lt;strong&gt;outside the system boundary entirely&lt;/strong&gt; — a layer that AI-assisted development fundamentally cannot corrupt.&lt;/p&gt;

&lt;p&gt;External automation tests interact with your backend the way a real client would. They don't care how the code is structured inside. They only ask one question:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Does the system still behave correctly?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This creates a clean divide between two things that should never be tightly coupled — &lt;strong&gt;system implementation&lt;/strong&gt; and &lt;strong&gt;system validation&lt;/strong&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  What's Inside the Full Article
&lt;/h2&gt;

&lt;p&gt;In the complete research article I cover:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;📖 A real-world case study of behavioral drift in a production backend&lt;/li&gt;
&lt;li&gt;🔍 Why conditional failures are the hardest to catch and the most costly&lt;/li&gt;
&lt;li&gt;🧪 The 4 layers of external automation testing every complex backend needs&lt;/li&gt;
&lt;li&gt;👥 How external automation improves team confidence and collaboration&lt;/li&gt;
&lt;li&gt;🗺️ A practical 6-step blueprint to build your own independent validation layer&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Read the Full Article on Medium
&lt;/h2&gt;

&lt;p&gt;This post is a glimpse. The full research article — with detailed case studies, diagrams, and a complete implementation blueprint — is published on Medium.&lt;/p&gt;

&lt;h3&gt;
  
  
  👉 &lt;a href="https://medium.com/@ashish.w.nadar/ai-backend-behavioral-drift-external-automation-testing-0e656bfc6410" rel="noopener noreferrer"&gt;Read the full article here&lt;/a&gt;
&lt;/h3&gt;




&lt;p&gt;&lt;em&gt;Ashish Nadar is a Scholar in Secure Cloud, Operational Excellence, and AI-Enabled Software researching how AI tooling reshapes the way backend systems are built, tested, and maintained.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Portfolio → &lt;a href="https://www.ashishnadar.com" rel="noopener noreferrer"&gt;ashishnadar.com&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>softwareengineering</category>
      <category>testing</category>
      <category>backend</category>
    </item>
  </channel>
</rss>
