<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Brandon Beam</title>
    <description>The latest articles on DEV Community by Brandon Beam (@brandon_beam_f4b2752055f4).</description>
    <link>https://dev.to/brandon_beam_f4b2752055f4</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3909669%2Fcf018c31-5dce-4500-bdd2-ede0d61c43bc.jpg</url>
      <title>DEV Community: Brandon Beam</title>
      <link>https://dev.to/brandon_beam_f4b2752055f4</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/brandon_beam_f4b2752055f4"/>
    <language>en</language>
    <item>
      <title>Gemma 4 Challenge: Full Stack Vibes, a public-good data refinery.</title>
      <dc:creator>Brandon Beam</dc:creator>
      <pubDate>Thu, 07 May 2026 20:56:50 +0000</pubDate>
      <link>https://dev.to/brandon_beam_f4b2752055f4/gemma-4-challenge-full-stack-vibes-a-public-good-data-refinery-2h5h</link>
      <guid>https://dev.to/brandon_beam_f4b2752055f4/gemma-4-challenge-full-stack-vibes-a-public-good-data-refinery-2h5h</guid>
      <description>&lt;p&gt;This is a submission for the &lt;strong&gt;Gemma 4 Challenge: Build with Gemma 4&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Built
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;FullStackVibes (FSV)&lt;/strong&gt; is a public-good context engineering commons. It provides verified, source-linked, and provenance-rich context artifacts via a &lt;strong&gt;Precision Bundle&lt;/strong&gt; retrieval API designed for small-model agents.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Thesis
&lt;/h3&gt;

&lt;p&gt;The post-LLM software lifecycle—where agents write code and "vibecoders" ship features—needs a shared, verified knowledge layer. Without one, every team wastes time rebuilding the same prompts and retrieval scaffolding in private. With FSV, that work compounds for everyone.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Motto:&lt;/strong&gt; &lt;em&gt;If it can be vibecoded, it must be documented.&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  How it Works
&lt;/h3&gt;

&lt;p&gt;FSV is composed of three core parts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;A Verified Corpus:&lt;/strong&gt; A collection of context artifacts (e.g., prompt-injection defense, Postgres migrations, HMAC signing) that are immutable, sha256-versioned, and human-verified.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Precision Bundle API:&lt;/strong&gt; A retrieval endpoint (&lt;code&gt;POST /api/v1/handshake&lt;/code&gt;) that returns context windows optimized for small-model consumption.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Gemma-4 Pipeline:&lt;/strong&gt; An inference engine that decomposes every submission into a structured, typed-window format.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Read access is free and unauthenticated. The API is the product, not a teaser.&lt;/p&gt;




&lt;h2&gt;
  
  
  Demo
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Live Site:&lt;/strong&gt; &lt;a href="https://fullstackvibes.com" rel="noopener noreferrer"&gt;https://fullstackvibes.com&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Search Engine Optimization:&lt;/strong&gt; Every artifact renders a full JSON-LD &lt;code&gt;@graph&lt;/code&gt;. This allows AI crawlers to see verified context without running JavaScript.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Try the API
&lt;/h3&gt;

&lt;p&gt;You can test the retrieval API right now with &lt;code&gt;curl&lt;/code&gt;:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;curl &lt;span class="nt"&gt;-X&lt;/span&gt; POST https://api.osenv.io/api/v1/handshake &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="nt"&gt;-H&lt;/span&gt; &lt;span class="s1"&gt;'Content-Type: application/json'&lt;/span&gt; &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="nt"&gt;-d&lt;/span&gt; &lt;span class="s1"&gt;'{
      "patternTags":  {"LIFECYCLE": ["hardening"]},
      "windowTypes":  ["CONSTRAINT", "ANTI_PATTERN"],
      "maxChars":     6000,
      "maxWindows":   12
    }'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;h2&gt;
  
  
  Code
&lt;/h2&gt;

&lt;p&gt;The corpus, API, and artifacts are open and inspectable.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Backend:&lt;/strong&gt; Rust (Axum) + PostgreSQL 16&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Frontend:&lt;/strong&gt; Server-side rendered HTML/JS&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Health Check:&lt;/strong&gt; &lt;a href="https://api.osenv.io/api/v1/health" rel="noopener noreferrer"&gt;https://api.osenv.io/api/v1/health&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;API Docs:&lt;/strong&gt; &lt;a href="https://fullstackvibes.com/docs/api/" rel="noopener noreferrer"&gt;https://fullstackvibes.com/docs/api/&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  How I Used Gemma 4
&lt;/h2&gt;

&lt;p&gt;The inference pipeline runs &lt;strong&gt;Gemma 4 E4B at 8-bit&lt;/strong&gt;, hosted locally. It handles five structured-output tasks for every submission:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;SLOP_DETECTION:&lt;/strong&gt; Filters out low-utility AI text before human review.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;QUALITY_REVIEW:&lt;/strong&gt; Scores submissions across multiple axes.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;RESOLVE_SPACES:&lt;/strong&gt; Automatically clusters artifacts into relevant use-cases.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;WINDOW_INDEX:&lt;/strong&gt; Breaks down bodies into types like &lt;code&gt;GOAL&lt;/code&gt;, &lt;code&gt;CONSTRAINT&lt;/code&gt;, and &lt;code&gt;ANTI_PATTERN&lt;/code&gt;.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;RESOLVE_TAGS:&lt;/strong&gt; Assigns tags for &lt;code&gt;AUDIENCE&lt;/code&gt;, &lt;code&gt;RISK&lt;/code&gt;, and &lt;code&gt;LIFECYCLE&lt;/code&gt;.&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Why E4B at 8-bit?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Public-Good Economics:&lt;/strong&gt; Local hosting makes marginal costs nearly zero (just electricity). This allows us to keep the corpus free for contributors without worrying about per-token API bills.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Sweet Spot:&lt;/strong&gt; At 8-bit, the model fits on commodity hardware with no noticeable loss in structured-output quality. It delivers reliable JSON shapes and consistent tagging.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Dogfood Loop:&lt;/strong&gt; Every inference call uses previously verified context windows in its system prompt. As the corpus grows, Gemma 4’s output quality compounds.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Gemma 4 E4B makes the "small-model retrieval" thesis financially and technically viable.&lt;/p&gt;

&lt;p&gt;Thanks, I hope you like my project.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>gemmachallenge</category>
      <category>gemma</category>
      <category>vibecached</category>
    </item>
    <item>
      <title>Can the 4b Gemma 4 at 16 bit really use CLI like a 120b?</title>
      <dc:creator>Brandon Beam</dc:creator>
      <pubDate>Sat, 02 May 2026 23:15:05 +0000</pubDate>
      <link>https://dev.to/brandon_beam_f4b2752055f4/can-the-4b-gemma-4-at-16-bit-really-use-cli-like-a-120b-3d1n</link>
      <guid>https://dev.to/brandon_beam_f4b2752055f4/can-the-4b-gemma-4-at-16-bit-really-use-cli-like-a-120b-3d1n</guid>
      <description></description>
      <category>ai</category>
      <category>cli</category>
      <category>google</category>
      <category>llm</category>
    </item>
  </channel>
</rss>
