<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: alethios000</title>
    <description>The latest articles on DEV Community by alethios000 (@alethios000).</description>
    <link>https://dev.to/alethios000</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3863971%2Ffa8724f6-9d35-4ceb-b844-3480d12a2c62.jpeg</url>
      <title>DEV Community: alethios000</title>
      <link>https://dev.to/alethios000</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/alethios000"/>
    <language>en</language>
    <item>
      <title>I built an open-source benchmark that scores AI agents, not models</title>
      <dc:creator>alethios000</dc:creator>
      <pubDate>Mon, 06 Apr 2026 13:15:48 +0000</pubDate>
      <link>https://dev.to/alethios000/i-built-an-open-source-benchmark-that-scores-ai-agents-not-models-36aa</link>
      <guid>https://dev.to/alethios000/i-built-an-open-source-benchmark-that-scores-ai-agents-not-models-36aa</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3wxpxqqbttb36cb4xura.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3wxpxqqbttb36cb4xura.png" alt=" " width="800" height="421"&gt;&lt;/a&gt;&lt;br&gt;
Two agents built on the same GPT-4o can have wildly different reliability. But every benchmark only evaluates the model.&lt;/p&gt;

&lt;p&gt;So I built Legit — an open-source platform that scores the agent as a whole.&lt;/p&gt;

&lt;h2&gt;
  
  
  How it works
&lt;/h2&gt;

&lt;p&gt;&lt;code&gt;pip install getlegit&lt;br&gt;
legit init --agent "MyBot" --endpoint "http://localhost:8000/run"&lt;br&gt;
legit run v1 --local&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;36 tasks across 6 categories (Research, Extract, Analyze, Code, Write, Operate). Two scoring layers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Layer 1: deterministic checks, runs locally, free&lt;/li&gt;
&lt;li&gt;Layer 2: 3 AI judges (Claude, GPT-4o, Gemini), median score&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Agents get an Elo rating and tier (Platinum/Gold/Silver/Bronze).&lt;/p&gt;

&lt;p&gt;Free, Apache 2.0.&lt;/p&gt;

&lt;p&gt;GitHub: &lt;a href="https://github.com/getlegitdev/legit" rel="noopener noreferrer"&gt;https://github.com/getlegitdev/legit&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Would love feedback on the scoring methodology!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>python</category>
      <category>agents</category>
    </item>
  </channel>
</rss>
