<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Jedsadakorn Suma</title>
    <description>The latest articles on DEV Community by Jedsadakorn Suma (@peachjed).</description>
    <link>https://dev.to/peachjed</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3883837%2F5f4a5c76-d329-4be9-bce1-e1134901d905.png</url>
      <title>DEV Community: Jedsadakorn Suma</title>
      <link>https://dev.to/peachjed</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/peachjed"/>
    <language>en</language>
    <item>
      <title>I built a deploy template pack for vibe-coded apps — here's what's inside and how I made it</title>
      <dc:creator>Jedsadakorn Suma</dc:creator>
      <pubDate>Fri, 17 Apr 2026 09:45:54 +0000</pubDate>
      <link>https://dev.to/peachjed/i-built-a-deploy-template-pack-for-vibe-coded-apps-heres-whats-inside-and-how-i-made-it-5aif</link>
      <guid>https://dev.to/peachjed/i-built-a-deploy-template-pack-for-vibe-coded-apps-heres-whats-inside-and-how-i-made-it-5aif</guid>
      <description>&lt;p&gt;I kept repeating the same painful cycle: build something in Cursor, then spend 2-3 hours copy-pasting Dockerfiles,&lt;br&gt;
  debugging nginx SSL configs, and fixing GitHub Actions YAML before I could actually ship.&lt;/p&gt;

&lt;p&gt;So I packaged everything into Vibe Deploy Kit.&lt;/p&gt;

&lt;p&gt;## What I built&lt;/p&gt;

&lt;p&gt;3 deployment template packs, each covering a different stack:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;next-production-kit&lt;/strong&gt; → Next.js 14+ to Vercel or self-hosted VPS&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;fastapi-deploy-kit&lt;/strong&gt; → FastAPI + Python to Railway (free tier)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;react-supabase-kit&lt;/strong&gt; → React + Vite + Supabase to Vercel&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;## How I built it&lt;/p&gt;

&lt;p&gt;Started by listing every file I was copy-pasting project to project. Turned out it was always the same ~5 files:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Dockerfile (multi-stage, non-root user, Alpine base)&lt;/li&gt;
&lt;li&gt;docker-compose.yml (app + nginx services, healthcheck)&lt;/li&gt;
&lt;li&gt;nginx.conf (rate limiting, SSL, security headers, gzip)&lt;/li&gt;
&lt;li&gt;GitHub Actions workflow (preview on PR, prod on push to main)&lt;/li&gt;
&lt;li&gt;.env.example (every required variable documented)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;## next-production-kit — the Dockerfile&lt;/p&gt;

&lt;p&gt;Multi-stage build keeps the final image small. Non-root user for security:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight docker"&gt;&lt;code&gt;  FROM node:20-alpine AS deps
  WORKDIR /app
  COPY package.json package-lock.json* pnpm-lock.yaml* yarn.lock* ./
  RUN \
    if [ -f package-lock.json ]; then npm ci; \
    elif [ -f pnpm-lock.yaml ]; then corepack enable pnpm &amp;amp;&amp;amp; pnpm i --frozen-lockfile; \
    else npm i; fi

  FROM node:20-alpine AS runner
  RUN addgroup --system --gid 1001 nodejs \
    &amp;amp;&amp;amp; adduser --system --uid 1001 nextjs
  COPY --from=builder --chown=nextjs:nodejs /app/.next/standalone ./
  USER nextjs
  CMD ["node", "server.js"]
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Requires &lt;code&gt;output: 'standalone'&lt;/code&gt; in next.config.js.&lt;/p&gt;

&lt;p&gt;## fastapi-deploy-kit — rate limiting&lt;/p&gt;

&lt;p&gt;The FastAPI kit uses slowapi so you don't get hammered on Railway's free tier:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;  &lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;slowapi&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;Limiter&lt;/span&gt;
  &lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;slowapi.util&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;get_remote_address&lt;/span&gt;

  &lt;span class="n"&gt;limiter&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nc"&gt;Limiter&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;key_func&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;get_remote_address&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

  &lt;span class="nd"&gt;@app.get&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;/api/hello&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
  &lt;span class="nd"&gt;@limiter.limit&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;30/minute&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
  &lt;span class="k"&gt;async&lt;/span&gt; &lt;span class="k"&gt;def&lt;/span&gt; &lt;span class="nf"&gt;hello&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;request&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;Request&lt;/span&gt;&lt;span class="p"&gt;):&lt;/span&gt;
      &lt;span class="k"&gt;return&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;message&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Hello&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;## react-supabase-kit — the client setup&lt;/p&gt;

&lt;p&gt;One thing vibe-coded apps almost never have is a proper ErrorBoundary. The kit includes one, plus a Supabase client&lt;br&gt;
  with auto-refresh:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight typescript"&gt;&lt;code&gt;  &lt;span class="k"&gt;export&lt;/span&gt; &lt;span class="kd"&gt;const&lt;/span&gt; &lt;span class="nx"&gt;supabase&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;createClient&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="nx"&gt;url&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="nx"&gt;anonKey&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
    &lt;span class="na"&gt;auth&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
      &lt;span class="na"&gt;autoRefreshToken&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="kc"&gt;true&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
      &lt;span class="na"&gt;persistSession&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="kc"&gt;true&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
      &lt;span class="na"&gt;detectSessionInUrl&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="kc"&gt;true&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="p"&gt;},&lt;/span&gt;
  &lt;span class="p"&gt;});&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;## nginx WebSocket config&lt;/p&gt;

&lt;p&gt;If you're self-hosting Next.js, you need this or hot reload breaks:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight nginx"&gt;&lt;code&gt;  &lt;span class="k"&gt;proxy_http_version&lt;/span&gt; &lt;span class="mf"&gt;1.1&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
  &lt;span class="k"&gt;proxy_set_header&lt;/span&gt; &lt;span class="s"&gt;Upgrade&lt;/span&gt; &lt;span class="nv"&gt;$http_upgrade&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
  &lt;span class="k"&gt;proxy_set_header&lt;/span&gt; &lt;span class="s"&gt;Connection&lt;/span&gt; &lt;span class="s"&gt;"upgrade"&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;## Tools used&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Docker + nginx for self-hosted path&lt;/li&gt;
&lt;li&gt;Vercel + Railway + Supabase for zero-cost cloud path&lt;/li&gt;
&lt;li&gt;GitHub Actions for CI/CD&lt;/li&gt;
&lt;li&gt;Gumroad for distribution&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;## Result&lt;/p&gt;

&lt;p&gt;A ZIP you drop into any project. Fill in .env, push to main → deployed.&lt;/p&gt;

&lt;p&gt;Available for $9: peachjed.gumroad.com/l/frosbp&lt;/p&gt;

</description>
      <category>docker</category>
      <category>nextjs</category>
      <category>devops</category>
      <category>webdev</category>
    </item>
    <item>
      <title>Self-host Open WebUI + Ollama in production — the config nobody writes about</title>
      <dc:creator>Jedsadakorn Suma</dc:creator>
      <pubDate>Fri, 17 Apr 2026 09:38:33 +0000</pubDate>
      <link>https://dev.to/peachjed/self-host-open-webui-ollama-in-production-the-config-nobody-writes-about-5cjf</link>
      <guid>https://dev.to/peachjed/self-host-open-webui-ollama-in-production-the-config-nobody-writes-about-5cjf</guid>
      <description>&lt;p&gt;Open WebUI's quickstart is great for local dev. One command, it's running. But putting it on a real server for a team&lt;br&gt;
  requires a lot more — SSL, auth lockdown, WebSocket proxy, backups.&lt;/p&gt;

&lt;p&gt;Here's the full production config I use.&lt;/p&gt;

&lt;p&gt;## The stack&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;ollama&lt;/strong&gt; — LLM runner&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;openwebui&lt;/strong&gt; — Chat UI&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;nginx&lt;/strong&gt; — Reverse proxy + SSL&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;## The nginx config that trips everyone up&lt;/p&gt;

&lt;p&gt;Open WebUI uses WebSockets for streaming. Without this, responses just hang:&lt;/p&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
nginx
  proxy_http_version 1.1;
  proxy_set_header Upgrade $http_upgrade;
  proxy_set_header Connection "upgrade";
  proxy_read_timeout 300s;

  Also set client_max_body_size 50M — users will upload documents and images.

  Lock down auth

  By default Open WebUI allows anyone to sign up. For a team setup:

  WEBUI_AUTH=true
  ENABLE_SIGNUP=false

  Now only the admin can create accounts. Add SMTP config to send email invites.

  GPU passthrough (NVIDIA)

  In docker-compose.yml, uncomment:

  deploy:
    resources:
      reservations:
        devices:
          - driver: nvidia
            count: all
            capabilities: [gpu]

  Then install nvidia-container-toolkit and restart. Without this, Ollama runs on CPU — still works but much slower.

  Automated backups

  docker run --rm \
    --volumes-from "$(docker compose ps -q openwebui)" \
    -v "$(pwd)/backups:/backup" \
    alpine \
    tar czf "/backup/openwebui_$(date +%Y%m%d).tar.gz" /app/backend/data

  Add to cron: 0 2 * * * /opt/openwebui/scripts/backup.sh

  Recommended free VPS

  Oracle Cloud Always Free — 4 vCPU / 24GB RAM ARM. Enough for llama3.2 (2B) with a small team.

  Full kit

  I packaged everything above into a ZIP — docker-compose, nginx config, backup + model scripts, .env.example.

  Gumroad: https://peachjed.gumroad.com/l/cazucc
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>selfhosted</category>
      <category>docker</category>
      <category>ollama</category>
      <category>opensource</category>
    </item>
    <item>
      <title>I built a runbook generator with FastAPI and Groq — here's how it works</title>
      <dc:creator>Jedsadakorn Suma</dc:creator>
      <pubDate>Fri, 17 Apr 2026 09:08:31 +0000</pubDate>
      <link>https://dev.to/peachjed/i-built-a-runbook-generator-with-fastapi-and-groq-heres-how-it-works-2go</link>
      <guid>https://dev.to/peachjed/i-built-a-runbook-generator-with-fastapi-and-groq-heres-how-it-works-2go</guid>
      <description>&lt;h2&gt;
  
  
  Every time there's an incident or a deployment, I need a runbook. The structure is always the same: Overview → Prerequisites → Procedure → Verification → Troubleshooting → Rollback.
&lt;/h2&gt;

&lt;p&gt;I got tired of filling in the same skeleton over and over, so I built &lt;strong&gt;Auto Runbook Generator&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What it does
&lt;/h2&gt;

&lt;p&gt;Paste a Kubernetes YAML, a docker-compose file, or just describe your system in plain English. Pick a runbook type (deployment, incident response, rollback, maintenance, onboarding). Get a complete Markdown runbook back in seconds.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Live demo:&lt;/strong&gt; &lt;a href="https://auto-runbook-gen-v0-1-0.onrender.com" rel="noopener noreferrer"&gt;https://auto-runbook-gen-v0-1-0.onrender.com&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How I built it
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Backend: FastAPI + AsyncGroq&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The core is a single &lt;code&gt;/api/generate&lt;/code&gt; endpoint. It takes the user's input and runbook type, builds a prompt, and calls Groq's API with llama-3.3-70b.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;response&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="k"&gt;await&lt;/span&gt; &lt;span class="nf"&gt;get_groq&lt;/span&gt;&lt;span class="p"&gt;().&lt;/span&gt;&lt;span class="n"&gt;chat&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;completions&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;create&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;model&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;llama-3.3-70b-versatile&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;messages&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;
        &lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;role&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;system&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;content&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;SYSTEM_PROMPT&lt;/span&gt;&lt;span class="p"&gt;},&lt;/span&gt;
        &lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;role&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;user&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;content&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;user_prompt&lt;/span&gt;&lt;span class="p"&gt;},&lt;/span&gt;
    &lt;span class="p"&gt;],&lt;/span&gt;
    &lt;span class="n"&gt;max_tokens&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mi"&gt;2048&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;temperature&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mf"&gt;0.3&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;I use &lt;code&gt;AsyncGroq&lt;/code&gt; (not &lt;code&gt;Groq&lt;/code&gt;) because FastAPI is async — calling a sync client inside an async endpoint blocks the event loop.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rate limiting: slowapi&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Free tier gets 10 generations/day per IP. Pro API key bypasses this.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="nd"&gt;@app.post&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;/api/generate&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="nd"&gt;@limiter.limit&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;10/day&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="k"&gt;async&lt;/span&gt; &lt;span class="k"&gt;def&lt;/span&gt; &lt;span class="nf"&gt;generate&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;request&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;Request&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;...):&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;The system prompt matters a lot&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The prompt enforces a consistent structure with 6 sections every time:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gh"&gt;# [Service Name] Runbook&lt;/span&gt;
&lt;span class="gu"&gt;## Overview&lt;/span&gt;
&lt;span class="gu"&gt;## Prerequisites&lt;/span&gt;
&lt;span class="gu"&gt;## Procedure&lt;/span&gt;
&lt;span class="gu"&gt;## Verification&lt;/span&gt;
&lt;span class="gu"&gt;## Troubleshooting&lt;/span&gt;
&lt;span class="gu"&gt;## Rollback&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Low temperature (0.3) keeps outputs consistent and factual rather than creative.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Frontend: zero dependencies&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Single HTML file, vanilla JS. &lt;code&gt;fetch()&lt;/code&gt; to call the API, &lt;code&gt;navigator.clipboard&lt;/code&gt; to copy the result. No React, no build step.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Deploy: Docker on Render&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight docker"&gt;&lt;code&gt;&lt;span class="k"&gt;FROM&lt;/span&gt;&lt;span class="s"&gt; python:3.12-slim&lt;/span&gt;
&lt;span class="k"&gt;WORKDIR&lt;/span&gt;&lt;span class="s"&gt; /app&lt;/span&gt;
&lt;span class="k"&gt;COPY&lt;/span&gt;&lt;span class="s"&gt; requirements.txt .&lt;/span&gt;
&lt;span class="k"&gt;RUN &lt;/span&gt;pip &lt;span class="nb"&gt;install&lt;/span&gt; &lt;span class="nt"&gt;--no-cache-dir&lt;/span&gt; &lt;span class="nt"&gt;-r&lt;/span&gt; requirements.txt
&lt;span class="k"&gt;COPY&lt;/span&gt;&lt;span class="s"&gt; . .&lt;/span&gt;
&lt;span class="k"&gt;CMD&lt;/span&gt;&lt;span class="s"&gt; ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000"]&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Results
&lt;/h2&gt;

&lt;p&gt;The output quality is good enough to use directly for most common cases. Kubernetes and Docker scenarios work especially well since the LLM has strong training data for those.&lt;/p&gt;

&lt;h2&gt;
  
  
  Source
&lt;/h2&gt;

&lt;p&gt;GitHub: &lt;a href="https://github.com/Jedsadakorn-Suma/auto-runbook-gen" rel="noopener noreferrer"&gt;https://github.com/Jedsadakorn-Suma/auto-runbook-gen&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Self-hostable (MIT). Needs a free Groq API key from console.groq.com.&lt;/p&gt;

</description>
      <category>devops</category>
      <category>python</category>
      <category>fastapi</category>
      <category>opensource</category>
    </item>
    <item>
      <title>I built 3 MCP servers so I can ask Claude about my DevOps stack</title>
      <dc:creator>Jedsadakorn Suma</dc:creator>
      <pubDate>Fri, 17 Apr 2026 07:01:04 +0000</pubDate>
      <link>https://dev.to/peachjed/i-built-3-mcp-servers-so-i-can-ask-claude-about-my-devops-stack-4c08</link>
      <guid>https://dev.to/peachjed/i-built-3-mcp-servers-so-i-can-ask-claude-about-my-devops-stack-4c08</guid>
      <description>&lt;p&gt;Every time something looked off in production, I'd switch between 4 tabs:&lt;br&gt;
  Prometheus → check metrics, kubectl → check pods, Grafana → check dashboards, terminal → check logs.&lt;/p&gt;

&lt;p&gt;So I built &lt;strong&gt;MCP DevOps Pack&lt;/strong&gt; — 3 MCP servers that let Claude Desktop talk to your infra directly.&lt;/p&gt;

&lt;p&gt;## What's included&lt;/p&gt;

&lt;p&gt;| Package | What it does |&lt;br&gt;
  |---------|-------------|&lt;br&gt;
  | &lt;code&gt;@peachjed/mcp-prometheus&lt;/code&gt; | PromQL queries, firing alerts, rule inspection |&lt;br&gt;
  | &lt;code&gt;@peachjed/mcp-kubernetes&lt;/code&gt; | List pods, get logs, describe resources, watch events |&lt;br&gt;
  | &lt;code&gt;@peachjed/mcp-grafana&lt;/code&gt; | Search dashboards, list datasources, check alert states |&lt;/p&gt;

&lt;p&gt;## Install&lt;/p&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
bash
  npm install -g @peachjed/mcp-prometheus @peachjed/mcp-kubernetes @peachjed/mcp-grafana

  Configure Claude Desktop

  Add to your claude_desktop_config.json:

  {
    "mcpServers": {
      "prometheus": {
        "command": "mcp-prometheus",
        "env": { "PROMETHEUS_URL": "http://localhost:9090" }
      },
      "kubernetes": {
        "command": "mcp-kubernetes"
      },
      "grafana": {
        "command": "mcp-grafana",
        "env": {
          "GRAFANA_URL": "http://localhost:3000",
          "GRAFANA_TOKEN": "your-token"
        }
      }
    }
  }

  What you can ask Claude

  - "What's the current CPU usage across all nodes?"
  - "Show me the last 50 lines from pod api-server-xyz in production"
  - "Are there any firing alerts right now?"
  - "List all dashboards in the Infrastructure folder"

  How it works

  Each server is a small TypeScript process that runs locally via stdio. Claude Desktop spawns it automatically when
  needed. The Kubernetes server uses your existing ~/.kube/config — no extra auth setup.

  Stack

  - TypeScript + @modelcontextprotocol/sdk
  - @kubernetes/client-node for the k8s server
  - Prometheus and Grafana via their HTTP APIs

  Source

  GitHub: https://github.com/Jedsadakorn-Suma/mcp-devops-pack

  npm: @peachjed/mcp-prometheus, @peachjed/mcp-kubernetes, @peachjed/mcp-grafana

  Feedback welcome — especially if you use a different observability stack.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>devops</category>
      <category>opensource</category>
      <category>claude</category>
      <category>mcp</category>
    </item>
  </channel>
</rss>
