<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Leonid Khomenko</title>
    <description>The latest articles on DEV Community by Leonid Khomenko (@leeaao).</description>
    <link>https://dev.to/leeaao</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2962971%2F449c118b-2b16-4d68-9828-4aa0d5deccf6.jpg</url>
      <title>DEV Community: Leonid Khomenko</title>
      <link>https://dev.to/leeaao</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/leeaao"/>
    <language>en</language>
    <item>
      <title>Grok 3 — Elon Musk’s AI, 2 Months Later</title>
      <dc:creator>Leonid Khomenko</dc:creator>
      <pubDate>Tue, 15 Apr 2025 11:12:06 +0000</pubDate>
      <link>https://dev.to/leeaao/grok-3-elon-musks-ai-2-months-later-47dc</link>
      <guid>https://dev.to/leeaao/grok-3-elon-musks-ai-2-months-later-47dc</guid>
      <description>&lt;p&gt;At the end of February, Elon rolled out his latest model. Of course, it was "the best in the world."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F64k064120v20q5tamkm8.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F64k064120v20q5tamkm8.webp" alt="bestElon" width="800" height="339"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Is it really a Smartest AI on Earth?&lt;/p&gt;

&lt;p&gt;As usual, Musk brought the hype train. But there wasn't much objective data at launch. xAI's short &lt;a&gt;blog post&lt;/a&gt; mentioned that it was still in beta and the models were actively training.&lt;/p&gt;

&lt;p&gt;They flashed some benchmarks showing Grok 3 ahead. However, they did not give access to the API. Which is important because independent benchmarks use it for evaluation.&lt;/p&gt;

&lt;p&gt;So, Elon claims Grok 3 is "scarily smart" and beats everything else. But the only ways to check were chatting with it yourself or looking at their benchmarks.&lt;/p&gt;

&lt;p&gt;And those benchmarks? Take a look:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqvu6ris31x3koxf1c3j1.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqvu6ris31x3koxf1c3j1.webp" alt="benchmarks" width="800" height="270"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;See that lighter area on the right? That's the boost Grok got by having way more compute power  (test-time compute) to get more consistent answers. It's not exactly a fair fight.&lt;/p&gt;

&lt;p&gt;You probably know AI models often give slightly different answers each time—sometimes better, sometimes worse. Most benchmarks ignore this variability, evaluating only the first response (pass@1). It’s simpler and matches how we actually use AI—we expect a good answer on the first try.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3zv9pbzx0tspdms5586h.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3zv9pbzx0tspdms5586h.webp" alt="cons64" width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;But Grok results were all shown using &lt;a href="mailto:cons@64"&gt;cons@64&lt;/a&gt;. Meaning, it got 64 tries for each question and picked the most common answer. Then, xAI compared that boosted score against the pass@1 scores of competitors.&lt;/p&gt;

&lt;p&gt;So on one hand, they claim it's a next-gen model. On the other, they're using pretty cheap tricks.&lt;/p&gt;

&lt;p&gt;To be fair, in such a competitive field, all labs bend the rules. They cherry-pick benchmarks or exclude stronger models from comparisons—but rarely as blatantly.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpvbvffsz5csndp0xm23s.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpvbvffsz5csndp0xm23s.webp" alt="internal" width="737" height="1200"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Okay, benchmarks aside. What are experienced users saying after actually using it? The general consensus is:&lt;/p&gt;

&lt;p&gt;The model is huge but hasn’t brought breakthroughs. It still hallucinates and tends toward overly long responses.&lt;/p&gt;

&lt;p&gt;Performance-wise, Grok 3 lands somewhere near the top OpenAI models, maybe a bit better than DeepSeek and Google's stuff at the time of release. &lt;/p&gt;

&lt;p&gt;However, two months later, Gemini 2.5, Claude 3.7, and the new GPT-4o arrived. We also finally got partial API access for Grok 3 and its mini version. Unfortunately, only the mini version received the thinking mode in API.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuyl5avhigd6hhzgwz5o2.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuyl5avhigd6hhzgwz5o2.webp" alt="api" width="800" height="201"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So today we know it's expensive and definitely not the absolute best.&lt;/p&gt;

&lt;p&gt;But hold on, there's still more to the story. &lt;/p&gt;

&lt;p&gt;The model is interesting and worth looking at. And you have to hand it to them, Elon and xAI jumped into the market quickly, becoming a key player in record time.&lt;/p&gt;

&lt;h2&gt;
  
  
  1 – The Hardware
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm5z4w4v65ff9qnjkre4c.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm5z4w4v65ff9qnjkre4c.webp" alt="colossus" width="800" height="300"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The big story here? &lt;/p&gt;

&lt;p&gt;In 2024, xAI built a massive compute cluster. We're talking 100,000 Nvidia H100 GPUs up and running in just 4 months. Then they doubled that to 200,000 cards in another 3 months. &lt;/p&gt;

&lt;p&gt;Nvidia's CEO, Jensen Huang, &lt;a&gt;mentioned&lt;/a&gt; this usually takes about 4 years.&lt;/p&gt;

&lt;p&gt;This was a massive engineering feat. And this time, no funny business—it's the largest data center in the world. Nobody else has managed to link up that many GPUs in one spot.&lt;/p&gt;

&lt;p&gt;Typically, such clusters are multiple regular data centers linked by costly Infiniband cables. During training, these centers need to swap tons of data constantly. If the connection is slow, those pricey GPUs sit idle, which is bad news.&lt;/p&gt;

&lt;p&gt;A typical data center might have 10,000-20,000 GPUs, sucking down 20-30 megawatts of power. &lt;a&gt;For example&lt;/a&gt;, Microsoft (for OpenAI) operates 100k GPUs network in Arizona, and Meta runs 128k.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fptibvo96fqt1x3iu928d.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fptibvo96fqt1x3iu928d.webp" alt="H" width="800" height="532"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;See the two H-shaped buildings? That's two standard Meta data centers next to each other.&lt;/p&gt;

&lt;p&gt;Power needs for top-tier clusters have exploded up to 10x since 2022. We're now talking around 150 MW per cluster. That's like powering a small city. This creates a huge load on regional power grids. In some places, it's actually cheaper to generate the power than to deliver it because there aren't enough power lines. &lt;/p&gt;

&lt;p&gt;So, Elon enters this market way behind. And... does the "Elon thing." Hate his tweets all you want, the man knows how to build factories like nobody else.&lt;/p&gt;

&lt;p&gt;He bought an old Electrolux factory in Memphis and decided to build one giant data center instead of a network like everyone else. &lt;/p&gt;

&lt;p&gt;Predictably, power became an issue.&lt;/p&gt;

&lt;p&gt;The factory only had 7 MW from the local grid—enough for only 4,000 GPUs. The local utility, Tennessee Valley Authority, promised another 50 MW, but not until August. And xAI's own 150 MW substation was still being built, not ready until year-end.&lt;/p&gt;

&lt;p&gt;But waiting isn’t Musk’s style.&lt;/p&gt;

&lt;p&gt;Dylan Patel (from Semianalysis) &lt;a&gt;spotted&lt;/a&gt; via satellite images that Elon just brought in 14 massive mobile diesel generators from VoltaGrid. Hooked them up to 4 mobile substations and powered the data center. Literally trucked in the electricity.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz699c1c9xvzanh4qzqhk.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz699c1c9xvzanh4qzqhk.webp" alt="volta" width="588" height="1005"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Patel mentioned they might have bought up 30% of the entire US market for these generators (though I couldn't find anything on that).&lt;/p&gt;

&lt;p&gt;Impressively, the data center also uses liquid cooling. Only Google has really done this at scale before. This is a big deal because the next generation of Nvidia chips, the Blackwell B200s, require liquid cooling. Everyone else will have to retrofit their existing data centers.&lt;/p&gt;

&lt;p&gt;You can check out the first few minutes of this video to see what it looks like inside. I got a chuckle out of how hyped the guy is about gray boxes and cables:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=Jf8EPSBZU7Y" rel="noopener noreferrer"&gt;https://www.youtube.com/watch?v=Jf8EPSBZU7Y&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It's seriously cool engineering—just look at the cable management. &lt;/p&gt;

&lt;p&gt;No one has done such massive work in so little time.&lt;/p&gt;

&lt;h2&gt;
  
  
  2 – Even More Hardware!
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd2mhn67bga4jrymxids.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd2mhn67bga4jrymxids.webp" alt="b100" width="800" height="546"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Elon says by summer 2025, they'll have a 300k GPU cluster with Blackwell B200 chips. Given Musk’s habit of exaggeration, let's say it's realistically somewhere between  200-400k new chips by the end of 2025. B200 is roughly 2.2 times better than H100 for model training (based on Nov 2024 estimates).&lt;/p&gt;

&lt;p&gt;Musk even plans to build a dedicated 2.2 GW power plant. That's more power than a medium-sized city consumes.&lt;/p&gt;

&lt;p&gt;And he's not alone—all the big players are doing something similar:&lt;/p&gt;

&lt;p&gt;Meta is building two gas plants in Louisiana.&lt;/p&gt;

&lt;p&gt;OpenAI/Microsoft is setting up something similar in Texas.&lt;/p&gt;

&lt;p&gt;Amazon and Google are also building gigawatt-scale data centers.&lt;/p&gt;

&lt;p&gt;Why not nuclear? It's got the power, but building a nuclear plant takes way too long. You can't just pop one up next to your data center in a year. Wind and solar farms plus batteries are promising, but they also take too long to deploy at the needed scale.&lt;/p&gt;

&lt;p&gt;As a result, both Microsoft and Meta have already had to backtrack on their green renewable energy promises. They broke their backs lifting Moloch to Heaven!&lt;/p&gt;

&lt;h2&gt;
  
  
  3 – Grok 3 is Huge
&lt;/h2&gt;

&lt;p&gt;So, Elon built this massive, expensive box. Now what?&lt;/p&gt;

&lt;p&gt;Estimates suggest Grok 2 trained on ~20k H100s, while Grok 3 used over 100k. For context, GPT-4 trained for about 90-100 days on ~25k older A100 chips, with H100 roughly 2.25x faster.&lt;/p&gt;

&lt;p&gt;Doing the math, Grok 2 got about twice the computing power thrown at it compared to GPT-4. And Grok 3 got five times more than Grok 2. Google's Gemini 2.0 probably used a similar amount of hardware (100k of their own TPUv6 chips), but the model itself is likely smaller.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsx16wym2szt74chhblp.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsx16wym2szt74chhblp.webp" alt="largescale" width="800" height="373"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Basically, the total &lt;a&gt;compute cost&lt;/a&gt; for Grok 3 is an order of magnitude (10 times!) higher than its closest competitor. Sadly, we don't have public data for GPT-4.5 or Gemini 2.5.&lt;/p&gt;

&lt;p&gt;So they poured insane amounts of resources into building this mega-cluster, and the resulting model is... just on par with the incumbents. Definitely not leagues better.&lt;/p&gt;

&lt;p&gt;It seems xAI's expertise in training still lags behind OpenAI, Google, or Anthropic. They essentially brute-forced their way into the top tier. No magic tricks shown, just: "If brute force isn't solving your problem, you aren't using enough of it."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguclmdsot9dz45lhxfxx.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguclmdsot9dz45lhxfxx.webp" alt="algorythms" width="800" height="476"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;But there's a catch with that approach. Epoch AI &lt;a&gt;estimates&lt;/a&gt; that over the last decade, algorithmic improvements accounted for about a third of the progress in model capabilities. The other two-thirds came from just throwing more hardware and data at bigger models.&lt;/p&gt;

&lt;p&gt;Brute force worked for Grok 3 this time, but costs will grow exponentially while delivering less and less improvement. And xAI need to catch up on the algorithm side. The good news  is that now they're seen as pushing the frontier, so it will likely be much easier to attract top talent.&lt;/p&gt;

&lt;h2&gt;
  
  
  4 – What's Good About Grok?
&lt;/h2&gt;

&lt;p&gt;1, It's completely free (probably until the full release). &lt;/p&gt;

&lt;p&gt;And without Anthropic's tight limits, DeepSeek's outages, or OpenAI's paid tiers. &lt;/p&gt;

&lt;p&gt;Even with all the new models dropped in the last couple of months, Grok is still holding its own near the top of the &lt;a&gt;Chatbot Arena&lt;/a&gt; leaderboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp05qzhqyz8zonwagkxwz.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp05qzhqyz8zonwagkxwz.webp" alt="arena" width="800" height="559"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We now also have an independent benchmarking by &lt;a&gt;EpochAI&lt;/a&gt;:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2kq74iov0nl58tslu09f.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2kq74iov0nl58tslu09f.webp" alt="gpqa" width="800" height="724"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And by &lt;a&gt;LiveBench&lt;/a&gt;:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqbp12ukb8twinag70luu.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqbp12ukb8twinag70luu.webp" alt="live" width="800" height="428"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2, Reasoning &amp;amp; Deep Research Mode&lt;/p&gt;

&lt;p&gt;Back in February, free Deep Research feature was mostly Perplexity exclusive. Now, Google and OpenAI offer some in a basic tier—maybe Grok pushed them? &lt;/p&gt;

&lt;p&gt;This mode automatically analyzes 30-100 links (Google might do more) in minutes and spits out a detailed (and bloated) summary that you just need to skim and fact-check. It's way easier than researching anything from scratch. I've found Grok's version works faster than the others, so I've started using it when I need to research something. Like, when buying a new headphones.&lt;/p&gt;

&lt;p&gt;3, Integration with X &lt;/p&gt;

&lt;p&gt;This could be its killer feature: semantic search not just for keywords, but for what you meant. You can also ask it to summarize posts on a topic to track trends. Or to find recent posts from a specific user.&lt;/p&gt;

&lt;p&gt;Twitter is the closest to a real-time information platform, so thats great. But so far Grok often lags, pulling data from the last couple of days instead. &lt;/p&gt;

&lt;p&gt;4, The Unfiltered Stuff&lt;/p&gt;

&lt;p&gt;And for the grand finale, the 18+ mode. Grok is notoriously easy to jailbreak without much effort. You can get it to do... well, whatever you might want, from flirty voices to questionable recipes. The voice mode examples are particularly wild.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://x.com/goodside/status/1893932239718691167" rel="noopener noreferrer"&gt;https://x.com/goodside/status/1893932239718691167&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Listen to the end, it's hilarious!&lt;/p&gt;

&lt;p&gt;Ironically, Grok itself doesn't seem to hold Musk (or Trump) in high regard. When this came out, xAI tried a fix—literally hardcoding a rule that Grok couldn't criticize Elon. When that blew up, they blamed a former OpenAI employee for “not fitting the company culture.” Super cringe. &lt;/p&gt;

&lt;p&gt;The real issue is that Grok's opinions are just a reflection of its training data (i.e., the internet), not some intentional bias. Trying to patch these views without messing up the whole model is hard.&lt;/p&gt;

&lt;h2&gt;
  
  
  5 - Should You Bother Trying It?
&lt;/h2&gt;

&lt;p&gt;Definitely try it, but as your second pilot.&lt;/p&gt;

&lt;p&gt;TLDR:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Cost way more to train than competitors' models.&lt;/li&gt;
&lt;li&gt;Despite that, performance is almost on par with the best.&lt;/li&gt;
&lt;li&gt;But it's super fast and free (for now).&lt;/li&gt;
&lt;li&gt;The Deep Research mode is genuinely useful—try it if you haven't.&lt;/li&gt;
&lt;li&gt;More prone to hallucinations and jumping to conclusions too fast.&lt;/li&gt;
&lt;li&gt;Responses are usually well-structured but often feels bloated.&lt;/li&gt;
&lt;li&gt;Unique access to Twitter data.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;xAI proved capable of building world-class infrastructure at unprecedented speed. But in actual AI capabilities, they're basically buying their way to the top with sheer compute power.&lt;/p&gt;

&lt;p&gt;This adds another strong player pressuring OpenAI, Google, and Anthropic, pushing the AI industry toward commoditization. Competition is heating up and the exclusivity of top-tier models is fading.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz2kpqct6fcwuscesisuu.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz2kpqct6fcwuscesisuu.webp" alt="neuralink" width="800" height="816"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Enjoyed this? Give an upvote or subscribe to &lt;a&gt;my newsletter&lt;/a&gt;.&lt;br&gt;
I'd appreciate it!&lt;/p&gt;

</description>
      <category>ai</category>
    </item>
    <item>
      <title>Too many AIs</title>
      <dc:creator>Leonid Khomenko</dc:creator>
      <pubDate>Fri, 21 Mar 2025 09:33:44 +0000</pubDate>
      <link>https://dev.to/leeaao/too-many-ais-24nb</link>
      <guid>https://dev.to/leeaao/too-many-ais-24nb</guid>
      <description>&lt;p&gt;Since early 2025, AI labs have flooded us with so many new models that I'm struggling to keep up.&lt;/p&gt;

&lt;p&gt;But trends says nobody cares! &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu49k4y5yxnuw9pb2yrir.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu49k4y5yxnuw9pb2yrir.png" alt="Image description" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;There is only ChatGPT.&lt;/p&gt;

&lt;p&gt;How so?&lt;/p&gt;

&lt;p&gt;The new models are awesome, but their naming is a complete mess. Plus, you can't even tell models apart by benchmarks anymore. Plain "this one's the best, everyone use it." doesn't work now.&lt;/p&gt;

&lt;p&gt;In short, there are many truly fantastic AI models on the market, but few people actually use them.&lt;/p&gt;

&lt;p&gt;And that's a shame!&lt;/p&gt;

&lt;p&gt;I'll try to make sense of the naming chaos, explain the benchmark crisis, and share tips on how to choose the right model for your needs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Too Many Models, Terrible Names
&lt;/h2&gt;

&lt;p&gt;Dario Amodei has long joked we might create AGI before we learn to name our models clearly. Google is traditionally leading the confusion game:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frakttuuqgel12r3lehse.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frakttuuqgel12r3lehse.png" alt="Image description" width="800" height="292"&gt;&lt;/a&gt;&lt;br&gt;
To be fair, it makes some sense. Each "base" model now has lots of updates. They're not always groundbreaking enough to justify each update as a new version. That's where all these prefixes come from.&lt;/p&gt;

&lt;p&gt;To simplify things, I put together a table of model types from major labs, stripping out all the unnecessary details.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7s7ij6jzn86xqu8of93o.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7s7ij6jzn86xqu8of93o.png" alt="Image description" width="800" height="299"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So what are these types of models?&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;There are huge, powerful &lt;strong&gt;base&lt;/strong&gt; models. They're impressive but slow and costly at scale.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;That's why we invented &lt;strong&gt;distillation&lt;/strong&gt;: take a base model, train a more compact model on its answers, and you get roughly the same capabilities, just faster and cheaper.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This is especially critical for &lt;strong&gt;reasoning&lt;/strong&gt; models. The best performers now follow multi-step reasoning chains—plan the solution, execute, and verify the outcome. Effective but pricey.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;There are also specialized models: for search, super-cheap ones for simple tasks, or models for specific fields like medicine and law. Plus a separate group for images, video, and audio. I didn't include all these to avoid confusion. I also deliberately ignored some other models and labs to keep it as simple as possible.&lt;/p&gt;

&lt;p&gt;Sometimes more details just make things worse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87a7cg1fs5gpwfqjvd6i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87a7cg1fs5gpwfqjvd6i.png" alt="Image description" width="800" height="850"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  All Models Are Basically Equal Now
&lt;/h2&gt;

&lt;p&gt;It's become tough to pick a clear winner. Andrej Karpathy recently called this a "evaluation crisis."&lt;/p&gt;

&lt;p&gt;&lt;em&gt;It's unclear which metrics to look at now. MMLU is outdated, SWE-Bench is too narrow. Chatbot Arena is so popular that labs have learned to "hack" it.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F043fjjhxek28u5gh8shf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F043fjjhxek28u5gh8shf.png" alt="Image description" width="800" height="848"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Currently, there are several ways to evaluate models:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Narrow benchmarks measure very specific skills, like Python coding or hallucination rates. But models are getting smarter and mastering more tasks, so you can't measure their level with just one metric anymore.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw6hq3zkwjqxrxtqyuyw7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw6hq3zkwjqxrxtqyuyw7.png" alt="Image description" width="800" height="410"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Comprehensive benchmarks try capturing multiple dimensions with loads of metrics. But comparing all these scores quickly becomes chaotic. Note that people try to factor multiple of these complex benchmarks. Five or ten at a time! One model wins here, another there—good luck making sense of it. LifeBench has 3 metrics within each category. And that's just one benchmark among dozens.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F84r4u8pw133t0jwgfa97.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F84r4u8pw133t0jwgfa97.png" alt="Image description" width="800" height="439"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Arena, where humans blindly compare model answers based on personal preferences. Models get an ELO rating, like chess players. Win more often, get higher ELO. But this was great until the models got too close to each other.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5l2nqypepzqqsxu8f096.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5l2nqypepzqqsxu8f096.png" alt="Image description" width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A 35-point difference means a model is better just 55% of the time.&lt;/p&gt;

&lt;p&gt;As in chess, the player with the lower ELO still has a good chance to win. Even with a 100-point gap, a "worse" model still outperforms in a third of cases.&lt;/p&gt;

&lt;p&gt;And again—some tasks are better solved by one model, others by another. Choose a model higher on the list and one of your 10 requests might be better. Which one and how much better? Who knows.&lt;/p&gt;

&lt;h2&gt;
  
  
  So, How Do You Choose?
&lt;/h2&gt;

&lt;p&gt;For lack of better options, Karpathy suggests relying on the vibe-check.&lt;/p&gt;

&lt;p&gt;Test the models yourself and see which one feels right. Sure, it's easy to fool yourself. It’s subjective and prone to bias—but it's practical.&lt;/p&gt;

&lt;p&gt;Here's my personal advice:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;If the task is new—open multiple tabs with different models and compare results. Trust your gut on which model requires less tweaking or edits.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;If the task is more familiar, use only your best model.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Forget about chasing benchmark numbers. Focus on the UX you like and prioritize the subscription you're already willing to pay for.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;If you still want numbers, try &lt;a href="https://livebench.ai/#/" rel="noopener noreferrer"&gt;https://livebench.ai/#/&lt;/a&gt;. The creators claim it fixes common benchmarking issues like hacking, obsolescence, narrowness and subjectivity.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;For product creators, here's a &lt;a href="https://github.com/huggingface/evaluation-guidebook/" rel="noopener noreferrer"&gt;great guide&lt;/a&gt; from HuggingFace on how to set up your own benchmark. &lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Meanwhile, if you've been waiting for a sign to try something other than ChatGPT, here it is:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://claude.ai/" rel="noopener noreferrer"&gt;https://claude.ai/&lt;/a&gt;&lt;br&gt;
&lt;a href="https://gemini.google.com/" rel="noopener noreferrer"&gt;https://gemini.google.com/&lt;/a&gt;&lt;br&gt;
&lt;a href="https://grok.com/" rel="noopener noreferrer"&gt;https://grok.com/&lt;/a&gt;&lt;br&gt;
&lt;a href="https://chat.deepseek.com/" rel="noopener noreferrer"&gt;https://chat.deepseek.com/&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/watch?v=dQw4w9WgXcQ" rel="noopener noreferrer"&gt;https://chat.openai.com/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is my first post here, but I have more writing on my &lt;a href="https://commonstragedy.substack.com/" rel="noopener noreferrer"&gt;Substack&lt;/a&gt;.&lt;br&gt;
I'd appreciate it if you subscribed—but honestly, the dev community seems pretty cool, so I plan to keep writing more stuff here. &lt;br&gt;
Maybe smth usefull on each of the models I mentioned.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>openai</category>
    </item>
  </channel>
</rss>
