<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Steriani Karamanlis</title>
    <description>The latest articles on DEV Community by Steriani Karamanlis (@steriani_karamanlis_ad61a).</description>
    <link>https://dev.to/steriani_karamanlis_ad61a</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3845016%2Fbd3790ee-06c4-417b-a347-a1aadecf8143.png</url>
      <title>DEV Community: Steriani Karamanlis</title>
      <link>https://dev.to/steriani_karamanlis_ad61a</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/steriani_karamanlis_ad61a"/>
    <language>en</language>
    <item>
      <title>OpenAI just raised $122B. Frontier inference pricing hasn't moved in 9 weeks</title>
      <dc:creator>Steriani Karamanlis</dc:creator>
      <pubDate>Fri, 03 Apr 2026 15:12:54 +0000</pubDate>
      <link>https://dev.to/steriani_karamanlis_ad61a/openai-just-raised-122b-frontier-inference-pricing-hasnt-moved-in-9-weeks-5oi</link>
      <guid>https://dev.to/steriani_karamanlis_ad61a/openai-just-raised-122b-frontier-inference-pricing-hasnt-moved-in-9-weeks-5oi</guid>
      <description>&lt;p&gt;OpenAI just closed the largest venture round in history at $852B valuation. Record capital, record confidence in AI's future.&lt;br&gt;
But here's what's interesting from a market pricing perspective. Frontier model pricing has been completely flat for 9 consecutive weeks. The benchmark sits at $0.005714 per 1K input tokens across top tier flagship models globally.&lt;br&gt;
At the same time the spread between frontier and budget models is 7.1x. That's a significant gap that's been holding steady.&lt;br&gt;
So the question the market is now asking is which way does this go from here. Does record capital give frontier labs room to hold pricing while budget models keep improving? Or does the competitive pressure eventually compress the premium?&lt;br&gt;
For teams building on top of inference at scale this dynamic matters a lot. The model selection decision isn't just a capability question anymore, it's a cost strategy question.&lt;br&gt;
Curious what others think. Are you seeing this pressure in your own stack decisions?&lt;br&gt;
We publish weekly inference pricing intelligence at a7om.com if you want the underlying data.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>infrastructure</category>
      <category>devops</category>
    </item>
    <item>
      <title>Query Live AI Inference Pricing with the ATOM MCP Server</title>
      <dc:creator>Steriani Karamanlis</dc:creator>
      <pubDate>Thu, 26 Mar 2026 20:48:24 +0000</pubDate>
      <link>https://dev.to/steriani_karamanlis_ad61a/query-live-ai-inference-pricing-with-the-atom-mcp-server-59ne</link>
      <guid>https://dev.to/steriani_karamanlis_ad61a/query-live-ai-inference-pricing-with-the-atom-mcp-server-59ne</guid>
      <description>&lt;p&gt;If you've ever tried to compare LLM pricing across vendors you know how painful it is. One charges per token, another per character, another per request. Cached input discounts exist but good luck finding them. Context window pricing is buried. And by the time you've normalized everything into a spreadsheet something changed on a pricing page and your numbers are stale.&lt;/p&gt;

&lt;p&gt;This is the problem ATOM was built to solve. It tracks 2,583 SKUs across 47 vendors, normalizes everything to a common unit, and exposes it all through an MCP server your agents can query directly.&lt;/p&gt;

&lt;p&gt;Here's how to set it up and what you can actually do with it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What MCP gives you here
&lt;/h2&gt;

&lt;p&gt;Model Context Protocol lets AI agents connect to external data sources through a standardized interface. Claude, Cursor, Windsurf and others support it natively.&lt;/p&gt;

&lt;p&gt;Instead of pasting a pricing table into your prompt and hoping it's current, you give your agent a live connection to the source. It queries, reasons, and acts on real numbers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Setting up the ATOM MCP server
&lt;/h2&gt;

&lt;p&gt;ATOM's server is published on npm, Smithery, and the official MCP registry.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Claude Desktop&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Add this to your &lt;code&gt;claude_desktop_config.json&lt;/code&gt; and restart:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"mcpServers"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="nl"&gt;"atom-pricing"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="nl"&gt;"command"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"npx"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="nl"&gt;"args"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="s2"&gt;"-y"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"atom-mcp-server"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Cursor or Windsurf&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Add the server endpoint in your MCP settings:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;https://atom-mcp-server-production.up.railway.app/mcp
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Any other MCP client&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The server supports both HTTP SSE and stdio transport. Run it locally via npx or point at the Railway endpoint above.&lt;/p&gt;

&lt;h2&gt;
  
  
  The tools
&lt;/h2&gt;

&lt;p&gt;The free tier includes 4 tools that give you macro market intelligence with no login required. MCP PRO ($49/mo) unlocks the remaining 4, which give you model-level and vendor-level detail.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Free tier&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;What it does&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;list_vendors&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;All 47 tracked vendors with type and region&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;get_kpis&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;6 live market KPIs updated weekly&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;get_index_benchmarks&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;14 AIPI price indexes by modality and tier&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;get_market_stats&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Aggregate supply and cost structure data&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;MCP PRO&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;What it does&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;search_models&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Filter by context size, tool support, modality, price&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;get_model_detail&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Full spec and pricing for a specific model&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;compare_prices&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Cross-vendor comparison for a model family&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;get_vendor_catalog&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Full SKU list for a specific vendor&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  What it looks like in practice
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Check what the market looks like right now (free)&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;get_kpis
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This week's numbers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Output tokens cost 3.84x more than input tokens on average&lt;/li&gt;
&lt;li&gt;Cached input saves 69.7% vs standard input pricing&lt;/li&gt;
&lt;li&gt;Open source models run 80% cheaper than closed source equivalents&lt;/li&gt;
&lt;li&gt;Only 20.3% of SKUs in the index offer cached pricing at all&lt;/li&gt;
&lt;li&gt;The price gap between small and large models in the same family is 4.8x&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These are median figures across all tracked SKUs, recalculated every Monday.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Find the cheapest model with 100K+ context and tool calling (PRO)&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight yaml"&gt;&lt;code&gt;&lt;span class="s"&gt;search_models&lt;/span&gt;
&lt;span class="na"&gt;context_window_min&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="m"&gt;100000&lt;/span&gt;
&lt;span class="na"&gt;tool_calling&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="kc"&gt;true&lt;/span&gt;
&lt;span class="na"&gt;sort_by&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;input_price_asc&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Returns model-level results with normalized per-token pricing across vendors. The spread between cheapest and most expensive for functionally similar models is typically over 30x. That difference compounds fast at any real usage volume.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Compare vendors for a specific model family (PRO)&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight yaml"&gt;&lt;code&gt;&lt;span class="s"&gt;compare_prices&lt;/span&gt;
&lt;span class="na"&gt;model_family&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Llama&lt;/span&gt;&lt;span class="nv"&gt; &lt;/span&gt;&lt;span class="s"&gt;3.3&lt;/span&gt;&lt;span class="nv"&gt; &lt;/span&gt;&lt;span class="s"&gt;70B"&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Returns every vendor offering that model with normalized pricing so you can make a direct comparison without doing any unit conversion yourself.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why this is useful for agent architecture
&lt;/h2&gt;

&lt;p&gt;If you're building anything that makes a lot of LLM calls, model routing based on cost and capability is a real decision you're making, consciously or not. The cheapest model that can handle a task should handle it.&lt;/p&gt;

&lt;p&gt;With ATOM connected your agent can check current prices before picking a model, catch when a vendor changes pricing, estimate the cost of a planned workload before running it, and compare vendors for a specific capability requirement. That reasoning used to mean a spreadsheet someone had to maintain. Now it's a tool call.&lt;/p&gt;

&lt;h2&gt;
  
  
  A note on the data
&lt;/h2&gt;

&lt;p&gt;ATOM uses a chained matched-model methodology, the same logic you'd apply to a commodity price index. Every SKU is normalized to a common unit, timestamped, and verified. The point of the methodology is to eliminate composition bias so week-over-week comparisons are actually meaningful and not just reflecting which vendors got added or dropped.&lt;/p&gt;

&lt;p&gt;Full methodology at a7om.com/methodology.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try it
&lt;/h2&gt;

&lt;p&gt;Run &lt;code&gt;npx atom-mcp-server&lt;/code&gt; or search "ATOM" on Smithery. Free tier covers 4 tools with no login. MCP PRO is at a7om.com/mcp.&lt;/p&gt;

&lt;p&gt;The inference market now has a benchmark. Might as well use it.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>mcp</category>
      <category>webdev</category>
    </item>
    <item>
      <title>"H, I just joined DEV. I've spent the past year building ATOM, a live pricing benchmark for AI inference. Tracks 2,500+ SKUs across 47 vendors. First article dropping this week on querying it via MCP. Follow along if that's relevant to what you build."</title>
      <dc:creator>Steriani Karamanlis</dc:creator>
      <pubDate>Thu, 26 Mar 2026 20:13:32 +0000</pubDate>
      <link>https://dev.to/steriani_karamanlis_ad61a/h-i-just-joined-dev-ive-spent-the-past-year-building-atom-a-live-pricing-benchmark-for-ai-2nfg</link>
      <guid>https://dev.to/steriani_karamanlis_ad61a/h-i-just-joined-dev-ive-spent-the-past-year-building-atom-a-live-pricing-benchmark-for-ai-2nfg</guid>
      <description></description>
    </item>
  </channel>
</rss>
