<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Swarit Shukla</title>
    <description>The latest articles on DEV Community by Swarit Shukla (@swaritshukla).</description>
    <link>https://dev.to/swaritshukla</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3874476%2F4d18f40b-9c14-4429-8ebc-716b824b1461.jpg</url>
      <title>DEV Community: Swarit Shukla</title>
      <link>https://dev.to/swaritshukla</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/swaritshukla"/>
    <language>en</language>
    <item>
      <title>Nvidia Vs Everyone: Why everyone is building their own accelerators</title>
      <dc:creator>Swarit Shukla</dc:creator>
      <pubDate>Mon, 04 May 2026 06:41:50 +0000</pubDate>
      <link>https://dev.to/swaritshukla/nvidia-vs-everyone-why-everyone-is-building-their-own-accelerators-5a8m</link>
      <guid>https://dev.to/swaritshukla/nvidia-vs-everyone-why-everyone-is-building-their-own-accelerators-5a8m</guid>
      <description>&lt;p&gt;There has been a race in the AI industry where companies are now preferring their own custom accelerators to train AI models and generate inference from them. But why? Why all of a sudden is everyone ditching Nvidia and either building their own accelerators or just buying from the ones that make them? For example, Anthropic uses Google’s TPUs (Tensor Processing Units), Amazon is offering their own Trainium chips, and OpenAI is hinting at making their own custom accelerators. The answer is fairly simple, and it’s not because they don’t want Nvidia to dominate over the market; it’s because their accelerator chips are faster than those Nvidia GPUs. So, let’s discuss this in detail.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fim19drnz2asqdgqefkx6.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fim19drnz2asqdgqefkx6.webp" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Why is everyone building their own accelerators?&lt;br&gt;
Nvidia has a reputation of making the best accelerated computing chips in the world, and for a long time everyone used them, but in recent years we are seeing a shift from GPUs to ASICs (Application-Specific Integrated Circuits).&lt;/p&gt;

&lt;p&gt;Once an AI model has been trained, running and constantly generating inference from it using GPUs is not very economical. The cloud service providers and hyperscalers are looking for custom chips that strip away the unnecessary components of GPUs, handling only a specific work, hence reducing the TCO (Total Cost of Ownership).&lt;/p&gt;

&lt;p&gt;The current data centers are power-hungry; they consume multiple gigawatts of power. Many CEOs, including Sam Altman, say that power is the ultimate bottleneck for AI. So to minimize power consumption, companies are turning to ASIC chips which increase performance per watt.&lt;/p&gt;

&lt;p&gt;Companies want control on their hardware; relying solely on Nvidia means fighting for the GPUs and waiting in the queue. By having their own chips, they are simply reducing their dependency on Nvidia.&lt;/p&gt;

&lt;p&gt;The difference between general purpose accelerators and application-specific accelerators&lt;br&gt;
You can say that AI models are just mathematical equations, so when we say transformer architecture, we are just referring to a sequence of mathematical operations that approximately looks like:&lt;/p&gt;

&lt;p&gt;Multiply matrices (weights and input)&lt;br&gt;
Add some numbers (bias)&lt;br&gt;
Run softmax (a mathematical function)&lt;br&gt;
Move to the next layer&lt;br&gt;
How a general purpose accelerator handles it:&lt;br&gt;
A general purpose accelerator doesn’t have any idea what a transformer architecture looks like; it just knows how to do basic maths. So it fetches the instructions to follow, it looks like:&lt;/p&gt;

&lt;p&gt;Check the memory for instructions (it says multiply)&lt;br&gt;
Get number a and b from the memory&lt;br&gt;
Send them to the calculator unit&lt;br&gt;
Save the answer back to the memory&lt;br&gt;
Check the memory for the next instruction&lt;br&gt;
Most of the energy is spent moving data between memory and compute, including reading instructions.&lt;/p&gt;

&lt;p&gt;How an ASIC handles it:&lt;br&gt;
We already know what maths is coming in the sequence so we don’t need to fetch instructions and move data in and out of the memory, so they take the physical logic gates that multiply matrices and connect their output to the logic gates that do softmax.&lt;/p&gt;

&lt;p&gt;In ASIC chips there is no instruction reading. The data enters and the laws of physics force it to go through a set sequence that generates the output.&lt;/p&gt;

&lt;p&gt;It’s like an automated car wash, you just take the car in and a clean car comes out. No interventions are being made, no directions are given, just the car goes in and comes out clean.&lt;/p&gt;

&lt;p&gt;How is Nvidia different from others?&lt;br&gt;
Nvidia’s biggest moat isn’t their chip designs, it’s “CUDA” and “SUPPLY CHAIN”.&lt;/p&gt;

&lt;p&gt;CUDA stands for Compute Unified Device Architecture. It’s a parallel computing platform that enables developers to use any CUDA-enabled GPU as a general purpose computing device. Millions of developers write software for CUDA; if you wrote a piece of software and it works on your Nvidia GPU, then it will surely work on any NVIDIA GPU.&lt;/p&gt;

&lt;p&gt;So if you are a startup writing code for Nvidia GPUs, you instantly unlock the deployment for every cloud service because every cloud service company in the world runs Nvidia GPUs.&lt;/p&gt;

&lt;p&gt;While watching the Jensen Huang podcast with Dwarkesh Patel, I realized that Nvidia has invested heavily to eradicate the bottlenecks. They determine bottlenecks in the supply chain prior, and go all in to make sure that it doesn’t bother them.&lt;/p&gt;

&lt;p&gt;They have invested heavily in CoWoS packaging technology, silicon photonics (interconnects using light), and HBM (High Bandwidth Memory). Nvidia handles the upstream very well so that it can stay ahead of the competition.&lt;/p&gt;

&lt;p&gt;Why are Nvidia’s chips still better than ASIC chips?&lt;br&gt;
When you compare an ASIC with accelerators, on paper surely you will see ASICs outperforming Nvidia’s accelerators, but in real life that’s not the case. Nvidia accelerators give the best performance per watt in today’s time.&lt;/p&gt;

&lt;p&gt;Nvidia doesn’t just sell the accelerators and rest, they have a lot of high-performing software engineers that continuously optimize the lowest-level code that runs on the accelerators. Nvidia helps the companies to optimize their code so they can run super fast on their GPUs.&lt;/p&gt;

&lt;p&gt;Jensen, while on the podcast with Dwarkesh Patel, gave a very good analogy. He said anyone can drive an F1 car, but to push it to its absolute limits you need to have a lot of expertise. He was referring to F1 cars as his accelerators here; they send their best software engineers to the AI labs and optimize their software so that it runs best on Nvidia hardware.&lt;/p&gt;

&lt;p&gt;Here is the most important point: the bottleneck is not the chip, it’s the network. Modern AI models are not trained and run using a single chip; they are trained and run on multiple chips working together. So it doesn’t matter if your chip is super fast because most of the time it will be sitting idle and waiting for the data to arrive. Nvidia recognized this bottleneck a long time ago and acquired Mellanox, a company that builds high-speed networking. Nvidia builds NVLink switches and InfiniBand networking fabrics that connect them.&lt;/p&gt;

&lt;p&gt;I would encourage you to think of Nvidia not as a chip design company but as a “DATACENTER MAKING COMPANY”.&lt;/p&gt;

&lt;p&gt;Why isn’t Nvidia making ASIC chips?&lt;br&gt;
So here is the real question: if NVIDIA can design the best chips and has got the best networks, then why aren’t they making ASIC chips themselves?&lt;/p&gt;

&lt;p&gt;To answer this question let me ask you a different question. We know that ASIC chips are hardwired to follow a specific sequence of maths and give the output, but what if the fundamental architecture behind the AI models (which is Transformers) changes? The answer is those ASIC chips instantly go from accelerators to paperweights; they will not be able to perform computation because the architecture has been changed and the chips are hardwired to perform computation for a specific architecture.&lt;/p&gt;

&lt;p&gt;The total addressable market (TAM) is bigger for general purpose accelerators. If NVIDIA builds ASICs, then the ones who will be buying them would be the companies who run Large Language Models. But by making general purpose accelerators, Nvidia can sell them to:&lt;/p&gt;

&lt;p&gt;LLM-centric companies for running AI models&lt;br&gt;
Medical companies for drug discovery by simulating molecular physics&lt;br&gt;
Tesla for training their self-driving cars&lt;br&gt;
The weather companies to predict the weather&lt;br&gt;
So why go for a small one when you can get a bigger piece of the pie?&lt;/p&gt;

&lt;p&gt;It’s not like Nvidia simply ignores this ASICs rise, they acknowledge it. If you study their GPUs closely you will find they have something called tensor cores and transformer engines; these sections are designed to make the Transformer-specific calculations fast, so NVIDIA gets an edge here.&lt;/p&gt;

&lt;p&gt;What if someone comes with something truly revolutionary? It’s simple, they acquire them. That’s what happened when the startup “GROQ” came up with LPUs (Language Processing Units) which reduced the cost of inference by a lot and sped up the time it took for inference; they didn’t waste a single second to acquire their core technology and talent.&lt;/p&gt;

&lt;p&gt;I hope this article gives you good information about the current landscape of AI hardware. Thanks for reading :)&lt;br&gt;
By Swarit Shukla&lt;/p&gt;

</description>
      <category>ai</category>
      <category>google</category>
      <category>claude</category>
      <category>nvidia</category>
    </item>
    <item>
      <title>The Elegance of MoE: How Gemma 4's 26B Model Runs Like a 4B Model</title>
      <dc:creator>Swarit Shukla</dc:creator>
      <pubDate>Sun, 12 Apr 2026 06:57:28 +0000</pubDate>
      <link>https://dev.to/swaritshukla/the-elegance-of-moe-how-gemma-4s-26b-model-runs-like-a-4b-model-4kl4</link>
      <guid>https://dev.to/swaritshukla/the-elegance-of-moe-how-gemma-4s-26b-model-runs-like-a-4b-model-4kl4</guid>
      <description>&lt;p&gt;Google recently dropped its new family of open-source AI models, Gemma 4, but the variant that truly captured my interest is Gemma-4-26B-A4B-IT. The question is: how can a 26 billion parameter model only activate 4 billion parameters at a time? This is where the elegance lies. By only activating 4 billion parameters, it reduces the cost of compute a lot. So what’s the magic behind this? It turns out it uses a clever architecture called MoE (Mixture of Experts) that lets the model choose experts, and hence it only activates 4 billion parameters, making it extremely fast and compute-efficient.&lt;/p&gt;

&lt;p&gt;A Mixture of Experts model is not a giant monolith. Internally, it is divided into experts (for example, 128). Experts specialize in different fields like coding, physics, calculus, and literature. So instead of using a giant neural network, it uses smaller expert neural networks. Note that these experts are not predefined—the neural network learns this itself during backpropagation.&lt;/p&gt;

&lt;p&gt;Dense models vs Mixture of Experts&lt;br&gt;
Traditional dense models differ from Mixture of Experts. In a dense model, each input token is fed to all of the parameters—but not in the case of MoE.&lt;/p&gt;

&lt;p&gt;MoE uses a router that assigns a token to only the top k experts (usually 2 or 8). The router (which is a neural network) takes in the token as input and then computes probabilities for all the experts. The top two with the highest probability are assigned that token.&lt;/p&gt;

&lt;p&gt;So at a time, only four billion parameters are activated, and the remaining 22 billion sit idle.&lt;/p&gt;

&lt;p&gt;The restaurant analogy&lt;br&gt;
Think of it this way—there are two restaurants:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Dense restaurant&lt;br&gt;
You place an order.&lt;br&gt;
In a dense restaurant, that order gets passed to every chef, and every chef works on it. It doesn’t matter if the order is for pasta—even the dessert chef will work on it. After every chef works on the dish, they create a delicious pasta.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;MoE restaurant&lt;br&gt;
This is where the router—the manager of experts—comes into play.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In an MoE, instead of the order directly going to the chefs, it first goes to the manager. The manager then decides which two chefs will work on the dish. If the dish ordered is pasta, then the two chefs working on it would be:&lt;/p&gt;

&lt;p&gt;The Vegetable Chef (Entremetier): Boils the starch (the pasta noodles)&lt;br&gt;
The Sauce Chef (Saucier): Cooks the hot, savory meat sauce to pour over the top&lt;br&gt;
Together, they create a delicious pasta without making all the chefs in the restaurant work on it. It’s as good as the one made by a dense restaurant, but with fewer chefs involved.&lt;/p&gt;

&lt;p&gt;(The idea for this analogy came from The Bear show—it’s an amazing show, by the way. Check it out.)&lt;/p&gt;

&lt;p&gt;Total vs Active parameters&lt;br&gt;
Total parameters – This represents the amount of diverse knowledge an LLM has. Let’s say a model has 128 experts and 26 billion parameters. Those 26 billion parameters are spread across 128 experts in their fields. Some are good at math, some at literature, and they might also go niche—like an expert in pop culture, movies, and music.&lt;/p&gt;

&lt;p&gt;Active parameters – This represents the compute cost of the model. So if a model has 26 billion parameters but only activates 4 billion at a time, the model’s compute cost and response time become that of a 4 billion parameter model.&lt;/p&gt;

&lt;p&gt;The vRAM twist&lt;br&gt;
Even though the model becomes extremely efficient at generating inference and reduces the compute cost significantly, there’s still the angle of vRAM.&lt;/p&gt;

&lt;p&gt;It doesn’t matter if the model activates only 4 billion parameters at a time—you still have to load the whole 26 billion parameters into your vRAM. So even though your low-end hardware can run the model efficiently, to load the model you will still need enough vRAM, which may force you to use a high-end system.&lt;/p&gt;

&lt;p&gt;It might give you fast responses and consume less energy, but you will still need a powerful device.&lt;/p&gt;

&lt;p&gt;An intuitive demonstration&lt;br&gt;
Let’s say you input the sequence: “Indian cuisine is very…“ and the LLM has to complete it.&lt;/p&gt;

&lt;p&gt;Input – The token “Indian” arrives&lt;br&gt;
Router’s evaluation – Based on mathematical evaluation of the token “Indian”, the router selects the top 2 experts, which could be the ones that specialize in geography (#34) and food (#87). (Modern LLMs consist of multiple layers stacked together, so the router assigns the token to different experts repeatedly as it goes deeper into the architecture.)&lt;br&gt;
Computation – Only the parameters in experts #34 and #87 get activated and used for computation, while the remaining parameters stay idle&lt;br&gt;
Repetition – The model repeats the process, but this time the router might choose completely different experts based on the next token&lt;br&gt;
The History&lt;br&gt;
It might seem like a very novel idea to most people, but the actual concept was introduced more than three decades ago, in 1991, by Robert A. Jacobs, Michael I. Jordan, Steven J. Nowlan, and the “Godfather of AI,” Geoffrey E. Hinton. The paper was titled “Adaptive Mixtures of Local Experts.”&lt;/p&gt;

&lt;p&gt;The modern implementation of this idea came in 2017. The paper, titled “Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer,” was written by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean (the Google Brain team). This paper further refined the idea by introducing the concept of sparsity—it forced the neural network to activate only a small number of parameters at a time, making them highly efficient.&lt;/p&gt;

&lt;p&gt;by Swarit Shukla&lt;/p&gt;

</description>
      <category>ai</category>
      <category>nlp</category>
      <category>mixtureofexperts</category>
      <category>gemini</category>
    </item>
  </channel>
</rss>
