<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Nirbhay Gautam</title>
    <description>The latest articles on DEV Community by Nirbhay Gautam (@nirbhay_gautam).</description>
    <link>https://dev.to/nirbhay_gautam</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3883897%2F9e2689ec-069b-4d38-b530-ced60fdb04d7.png</url>
      <title>DEV Community: Nirbhay Gautam</title>
      <link>https://dev.to/nirbhay_gautam</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/nirbhay_gautam"/>
    <language>en</language>
    <item>
      <title>Gemma 4 Has Four Models. Here's Which One You Actually Need</title>
      <dc:creator>Nirbhay Gautam</dc:creator>
      <pubDate>Fri, 08 May 2026 10:58:32 +0000</pubDate>
      <link>https://dev.to/nirbhay_gautam/gemma-4-has-four-models-heres-which-one-you-actually-need-1057</link>
      <guid>https://dev.to/nirbhay_gautam/gemma-4-has-four-models-heres-which-one-you-actually-need-1057</guid>
      <description>&lt;p&gt;Google called it one launch. It's not.&lt;/p&gt;

&lt;p&gt;Gemma 4 is four completely different models with different architectures, different hardware requirements, and different use cases — packaged under one name that makes it sound like a single thing. If you read the announcement and walked away confused about what to actually download, that's not on you. That's the naming.&lt;/p&gt;

&lt;p&gt;I've been building with local AI for a while — I recently built a RAG system using Llama 3.2 running locally via Ollama, and the hardware reality of running LLMs on a regular laptop is something I've dealt with firsthand. So let me break this down practically, not theoretically.&lt;/p&gt;




&lt;h2&gt;
  
  
  First: What "E" and "A" Actually Mean
&lt;/h2&gt;

&lt;p&gt;The naming convention is doing a lot of work here and Google doesn't explain it upfront.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;E2B and E4B&lt;/strong&gt; — the "E" stands for &lt;em&gt;effective&lt;/em&gt; parameters. These are not 2B and 4B parameter models in the traditional sense. They use Per-Layer Embeddings (PLE) to pack more capability into fewer parameters. Think of it as parameter efficiency — more intelligence per byte than the raw number suggests.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;26B A4B&lt;/strong&gt; — the "A" stands for &lt;em&gt;active&lt;/em&gt; parameters. This is a Mixture-of-Experts (MoE) model with 26B total parameters, but during inference, only a 4B subset activates per token — making it run almost as fast as a 4B-parameter model. You get the quality of a large model at the speed of a small one.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;31B Dense&lt;/strong&gt; — no tricks. Every token touches all 31 billion parameters. Slower, heavier, but the most predictable behavior.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Four Models, Plainly
&lt;/h2&gt;

&lt;h3&gt;
  
  
  E2B — For the Edge
&lt;/h3&gt;

&lt;p&gt;Google's own tests show Gemma 4 E2B running on a Raspberry Pi 5 at around 7.6 tokens per second — slow but functional for edge agent workflows.&lt;/p&gt;

&lt;p&gt;If you're building something that needs to run on a phone, a microcontroller, or offline hardware with no GPU — this is your only option in the family. It supports audio natively, which the larger models don't. Context window tops out at 128K.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Who it's for:&lt;/strong&gt; IoT projects, on-device apps, offline deployments, Raspberry Pi builds.&lt;br&gt;
&lt;strong&gt;Who it's NOT for:&lt;/strong&gt; Anyone who wants quality answers on a regular laptop.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  E4B — The Practical Daily Driver
&lt;/h3&gt;

&lt;p&gt;E4B runs comfortably on any modern laptop — Mac, Windows, Linux — and delivers surprisingly good quality for its size.&lt;/p&gt;

&lt;p&gt;This is the model most developers should start with. It handles image input, audio, and text. It's fast enough for interactive use and doesn't require a dedicated GPU. Context window is 128K.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Who it's for:&lt;/strong&gt; Developers on regular laptops, multimodal projects that need audio, quick prototyping.&lt;br&gt;
&lt;strong&gt;Who it's NOT for:&lt;/strong&gt; Tasks requiring deep reasoning or complex long-document analysis.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  26B A4B — The Hidden Best Value
&lt;/h3&gt;

&lt;p&gt;This one is the most interesting in the lineup and the least talked about.&lt;/p&gt;

&lt;p&gt;The 26B A4B achieves roughly &lt;strong&gt;97% of the dense 31B model's quality&lt;/strong&gt; while activating only 3.8B parameters per token — about 8x less compute per inference step. On the LMArena leaderboard it scores 1441 Elo versus 1452 for the 31B — a gap that's invisible in most real-world tasks.&lt;/p&gt;

&lt;p&gt;If you have a machine with 16GB+ RAM and a decent GPU, or Apple Silicon, this is arguably the best model in the whole lineup. You get near-31B quality at a fraction of the compute cost. Context window extends to 256K here.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Who it's for:&lt;/strong&gt; Developers with a decent machine who want maximum quality-per-compute, agentic workflows, long-document tasks.&lt;br&gt;
&lt;strong&gt;Who it's NOT for:&lt;/strong&gt; Low-spec machines, anyone without at least 16GB RAM.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  31B Dense — Maximum Quality, Maximum Cost
&lt;/h3&gt;

&lt;p&gt;The 31B model currently ranks as the &lt;strong&gt;#3 open model in the world&lt;/strong&gt; on the Arena AI text leaderboard. Every token touches all 31 billion parameters. No shortcuts.&lt;/p&gt;

&lt;p&gt;It's slower than the 26B A4B for inference, but it's the better candidate if you want to fine-tune — dense architecture means cleaner gradient flow during training. Context window is 256K and it actually works: the 31B went from 13.5% to 66.4% on multi-needle retrieval tests, meaning the model can actually find and reason over information buried deep in a long document.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Who it's for:&lt;/strong&gt; Server deployments, fine-tuning projects, maximum quality use cases, cloud inference.&lt;br&gt;
&lt;strong&gt;Who it's NOT for:&lt;/strong&gt; Anyone running locally without a workstation-grade GPU.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  The Hardware Reality Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;Here's my honest take as someone who's actually tried running LLMs locally on consumer hardware:&lt;/p&gt;

&lt;p&gt;A few weeks ago I built a job market Q&amp;amp;A system using Llama 3.2 running locally via Ollama. The setup worked — but every response took 10-15 seconds on my CPU, and I spent more time watching a blinking cursor than actually using the thing.&lt;br&gt;
I stuck with local anyway, not because it was convenient, but because the alternative was sending job description data and user queries to an external API I don't control. For a portfolio project that's fine. For anything with real user data, that tradeoff stops being theoretical.&lt;br&gt;
And that's the honest hardware reality nobody talks about: the gap between "this model CAN run on your laptop" and "this model runs well on your laptop" is real and wide. &lt;br&gt;
E2B and E4B are the only Gemma 4 models most people can realistically run locally without a dedicated GPU. The 26B A4B and 31B are cloud or workstation territory for most developers.&lt;/p&gt;

&lt;p&gt;That experience is what made Gemma 4's range genuinely interesting to me — not the benchmarks, but the fact that someone with a decent laptop can now run a near-31B quality model locally without a GPU, or fall back to the 31B on OpenRouter's free tier without sacrificing open-weight guarantees. The hardware ceiling is still real.&lt;/p&gt;




&lt;h2&gt;
  
  
  Quick Decision Guide
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Your situation&lt;/th&gt;
&lt;th&gt;Use this&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Raspberry Pi / phone / IoT&lt;/td&gt;
&lt;td&gt;E2B&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Regular laptop, need audio&lt;/td&gt;
&lt;td&gt;E4B&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;16GB RAM + decent GPU&lt;/td&gt;
&lt;td&gt;26B A4B&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Server / fine-tuning / max quality&lt;/td&gt;
&lt;td&gt;31B Dense&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;No GPU, want best quality&lt;/td&gt;
&lt;td&gt;31B via OpenRouter (free)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;




&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;Gemma 4 is genuinely impressive — not because any single model is revolutionary, but because the family covers the full deployment spectrum from a Raspberry Pi to a workstation under one open license. That's rare.&lt;/p&gt;

&lt;p&gt;But &lt;strong&gt;"Gemma 4" is not one thing.&lt;/strong&gt; Pick the right model for your hardware, your use case, and your deployment target. The name is marketing. The specs are what matter.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>gemmachallenge</category>
      <category>gemma</category>
    </item>
    <item>
      <title>I'm a Final-Year CS Student — And I'm Done Letting AI Tools Own My Data</title>
      <dc:creator>Nirbhay Gautam</dc:creator>
      <pubDate>Fri, 17 Apr 2026 09:08:58 +0000</pubDate>
      <link>https://dev.to/nirbhay_gautam/im-a-final-year-cs-student-and-im-done-letting-ai-tools-own-my-data-556k</link>
      <guid>https://dev.to/nirbhay_gautam/im-a-final-year-cs-student-and-im-done-letting-ai-tools-own-my-data-556k</guid>
      <description>&lt;p&gt;There's a particular kind of restlessness that comes from being a CS student in 2026.&lt;br&gt;
It's not the assignments, or the deadlines, or even the exams. It's the feeling that by the time you've shipped something, the landscape has already shifted. Every week there's a new model, a new tool, a new framework that everyone on the internet insists will change everything. I've genuinely lost count of how many times I've read the words "the future is here" in the last six months alone.&lt;br&gt;
So when OpenClaw started showing up everywhere — 20,000 GitHub stars in a day, people on every forum talking about their "personal AI agent" like it was a houseplant they were proud of — my first instinct was to actually look into it. I'd learned to tell the difference between hype and something worth digging into. This felt like the latter.&lt;br&gt;
I was right.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why I Actually Tried It&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;It wasn't the star count that got me. It was one specific thing I kept reading about: you run it yourself. On your own machine. Your data doesn't disappear into some company's server farm. Your conversations, your memory, your context — it all lives in plain text files that you can open, read, and edit like any other document.&lt;/p&gt;

&lt;p&gt;That mattered to me more than I expected it to.&lt;/p&gt;

&lt;p&gt;As someone who's spent four years studying how these systems work under the hood, I've grown increasingly uncomfortable with how much of my digital life runs inside black boxes I'm not allowed to inspect. Most AI tools are designed to keep you dependent and ignorant — not out of malice, but because that's the model. OpenClaw felt like a deliberate rejection of that. Not a subscription. Not a walled garden. A tool that actually respects the intelligence of the person using it.&lt;/p&gt;

&lt;p&gt;The second pull was automation. I have a full plate — final year coursework, projects, part-time commitments, and a group chat that never sleeps. The idea that something could handle the repetitive edges of that without me having to babysit it wasn't a luxury. It was just practical.&lt;br&gt;
So I set it up. It took an afternoon, a couple of wrong turns, and one moment where I was fairly sure I'd misconfigured something permanently. But I got there.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The First Week: Honest Notes from the Trenches&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The first thing OpenClaw did that genuinely impressed me was remember something I told it on day one and surface it — unprompted — three days later in a completely different context. I already understood the mechanism: daily logs for short-term context, a long-term MEMORY.md file for the important stuff, all plain markdown. But seeing it work in practice still landed differently than reading about it. It felt less like a chatbot and more like a system that was actually tracking state in a meaningful way.&lt;/p&gt;

&lt;p&gt;The automation side was where things got interesting from an engineering perspective. The first workflow I built was simple — summarising a long document and routing the output somewhere useful. The setup took longer than the time it saved, at first. But that's always how it goes when you're building infrastructure rather than just using it. Once the pattern clicks, everything after it gets faster.&lt;/p&gt;

&lt;p&gt;What I didn't expect was how readable the internals would be. The memory files, the logs, the configuration — all of it is plain text you can inspect, version control, and reason about. For someone who's used to digging into source code to understand what a system is actually doing, that's not a small thing. It meant I could debug it like any other software rather than submitting a support ticket and hoping for the best.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What It Gets Right&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most AI tools are built on the assumption that you should trust them completely and ask no questions. You send input, you receive output, and the gap between the two is none of your business.&lt;/p&gt;

&lt;p&gt;OpenClaw makes the opposite bet. The architecture is transparent by design. Your memory is in files you own. Your logs are yours to read. Your configuration is yours to modify. It's a system that treats you as someone capable of understanding what's happening — because you are.&lt;/p&gt;

&lt;p&gt;That transparency has a compounding effect. Every time something didn't behave as expected and I had to dig into why, I came out with a clearer mental model of how AI agents actually work — the execution loops, the tool calls, the context management. That's knowledge that transfers. It doesn't matter what the next popular agent framework is; the underlying concepts are the same.&lt;/p&gt;

&lt;p&gt;The other thing it gets right is low friction adoption. You don't need a new app or a new habit. You can interface with it through messaging platforms you already use. The best tools are the ones that fit into your life rather than demanding you reorganize around them. At its best, OpenClaw disappears into your existing workflow.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;On the Pace of All This&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I want to be honest about something: the pace of this space is genuinely wild — even when you're close to it. OpenClaw went from zero to the most-starred project in GitHub history in months. Its creator got hired by OpenAI mid-project. Serious security vulnerabilities, a thriving skill marketplace, a conference, major players building on top of it. All while I was finishing coursework.&lt;/p&gt;

&lt;p&gt;Being a CS student doesn't make you immune to that pace — if anything, it makes you more aware of how much is happening simultaneously and how hard it is to separate signal from noise. The temptation is to try to follow everything. I've learned that's the wrong move. Better to go deep on something real than to skim the surface of everything.&lt;br&gt;
OpenClaw has become that something for me — not because it's the most polished tool out there, but because it's open, it's honest about what it is, and working with it has been more educational than most things I've read about AI agents this year.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Comes Next&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I'm still early with it. There's a lot I haven't touched — deeper skill integrations, multi-agent setups, building on top of the API layer. But I'm not in a rush.&lt;/p&gt;

&lt;p&gt;That might be the most useful reframe OpenClaw gave me: the goal isn't to keep up with everything. It's to actually build something, understand it properly, and carry that understanding forward regardless of what the landscape looks like next month.&lt;/p&gt;

&lt;p&gt;If you're a developer who's been watching the AI agent space from a distance, waiting for something worth getting into — this is worth your time. Not because it'll automate your life overnight. But because understanding it changes how you think about everything else in this space.&lt;/p&gt;

&lt;p&gt;And right now, that's the most useful thing a tool can do.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>openclawchallenge</category>
    </item>
    <item>
      <title>I'm a Final-Year CS Student — And I'm Done Letting AI Tools Own My Data</title>
      <dc:creator>Nirbhay Gautam</dc:creator>
      <pubDate>Fri, 17 Apr 2026 08:57:32 +0000</pubDate>
      <link>https://dev.to/nirbhay_gautam/im-a-final-year-cs-student-and-im-done-letting-ai-tools-own-my-data-2adn</link>
      <guid>https://dev.to/nirbhay_gautam/im-a-final-year-cs-student-and-im-done-letting-ai-tools-own-my-data-2adn</guid>
      <description>&lt;p&gt;There's a particular kind of restlessness that comes from being a CS student in 2026.&lt;br&gt;
It's not the assignments, or the deadlines, or even the exams. It's the feeling that by the time you've shipped something, the landscape has already shifted. Every week there's a new model, a new tool, a new framework that everyone on the internet insists will change everything. I've genuinely lost count of how many times I've read the words "the future is here" in the last six months alone.&lt;br&gt;
So when OpenClaw started showing up everywhere — 20,000 GitHub stars in a day, people on every forum talking about their "personal AI agent" like it was a houseplant they were proud of — my first instinct was to actually look into it. I'd learned to tell the difference between hype and something worth digging into. This felt like the latter.&lt;br&gt;
I was right.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why I Actually Tried It&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;It wasn't the star count that got me. It was one specific thing I kept reading about: you run it yourself. On your own machine. Your data doesn't disappear into some company's server farm. Your conversations, your memory, your context — it all lives in plain text files that you can open, read, and edit like any other document.&lt;/p&gt;

&lt;p&gt;That mattered to me more than I expected it to.&lt;/p&gt;

&lt;p&gt;As someone who's spent four years studying how these systems work under the hood, I've grown increasingly uncomfortable with how much of my digital life runs inside black boxes I'm not allowed to inspect. Most AI tools are designed to keep you dependent and ignorant — not out of malice, but because that's the model. OpenClaw felt like a deliberate rejection of that. Not a subscription. Not a walled garden. A tool that actually respects the intelligence of the person using it.&lt;/p&gt;

&lt;p&gt;The second pull was automation. I have a full plate — final year coursework, projects, part-time commitments, and a group chat that never sleeps. The idea that something could handle the repetitive edges of that without me having to babysit it wasn't a luxury. It was just practical.&lt;br&gt;
So I set it up. It took an afternoon, a couple of wrong turns, and one moment where I was fairly sure I'd misconfigured something permanently. But I got there.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The First Week: Honest Notes from the Trenches&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The first thing OpenClaw did that genuinely impressed me was remember something I told it on day one and surface it — unprompted — three days later in a completely different context. I already understood the mechanism: daily logs for short-term context, a long-term MEMORY.md file for the important stuff, all plain markdown. But seeing it work in practice still landed differently than reading about it. It felt less like a chatbot and more like a system that was actually tracking state in a meaningful way.&lt;/p&gt;

&lt;p&gt;The automation side was where things got interesting from an engineering perspective. The first workflow I built was simple — summarising a long document and routing the output somewhere useful. The setup took longer than the time it saved, at first. But that's always how it goes when you're building infrastructure rather than just using it. Once the pattern clicks, everything after it gets faster.&lt;/p&gt;

&lt;p&gt;What I didn't expect was how readable the internals would be. The memory files, the logs, the configuration — all of it is plain text you can inspect, version control, and reason about. For someone who's used to digging into source code to understand what a system is actually doing, that's not a small thing. It meant I could debug it like any other software rather than submitting a support ticket and hoping for the best.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What It Gets Right&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most AI tools are built on the assumption that you should trust them completely and ask no questions. You send input, you receive output, and the gap between the two is none of your business.&lt;/p&gt;

&lt;p&gt;OpenClaw makes the opposite bet. The architecture is transparent by design. Your memory is in files you own. Your logs are yours to read. Your configuration is yours to modify. It's a system that treats you as someone capable of understanding what's happening — because you are.&lt;/p&gt;

&lt;p&gt;That transparency has a compounding effect. Every time something didn't behave as expected and I had to dig into why, I came out with a clearer mental model of how AI agents actually work — the execution loops, the tool calls, the context management. That's knowledge that transfers. It doesn't matter what the next popular agent framework is; the underlying concepts are the same.&lt;/p&gt;

&lt;p&gt;The other thing it gets right is low friction adoption. You don't need a new app or a new habit. You can interface with it through messaging platforms you already use. The best tools are the ones that fit into your life rather than demanding you reorganize around them. At its best, OpenClaw disappears into your existing workflow.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;On the Pace of All This&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I want to be honest about something: the pace of this space is genuinely wild — even when you're close to it. OpenClaw went from zero to the most-starred project in GitHub history in months. Its creator got hired by OpenAI mid-project. Serious security vulnerabilities, a thriving skill marketplace, a conference, major players building on top of it. All while I was finishing coursework.&lt;/p&gt;

&lt;p&gt;Being a CS student doesn't make you immune to that pace — if anything, it makes you more aware of how much is happening simultaneously and how hard it is to separate signal from noise. The temptation is to try to follow everything. I've learned that's the wrong move. Better to go deep on something real than to skim the surface of everything.&lt;br&gt;
OpenClaw has become that something for me — not because it's the most polished tool out there, but because it's open, it's honest about what it is, and working with it has been more educational than most things I've read about AI agents this year.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Comes Next&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I'm still early with it. There's a lot I haven't touched — deeper skill integrations, multi-agent setups, building on top of the API layer. But I'm not in a rush.&lt;/p&gt;

&lt;p&gt;That might be the most useful reframe OpenClaw gave me: the goal isn't to keep up with everything. It's to actually build something, understand it properly, and carry that understanding forward regardless of what the landscape looks like next month.&lt;/p&gt;

&lt;p&gt;If you're a developer who's been watching the AI agent space from a distance, waiting for something worth getting into — this is worth your time. Not because it'll automate your life overnight. But because understanding it changes how you think about everything else in this space.&lt;/p&gt;

&lt;p&gt;And right now, that's the most useful thing a tool can do.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>openclawchallenge</category>
    </item>
  </channel>
</rss>
