<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Steven Hooley</title>
    <description>The latest articles on DEV Community by Steven Hooley (@sbhooley).</description>
    <link>https://dev.to/sbhooley</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3872545%2F2db27926-d232-4ca6-9ed0-1d49689cb87e.jpg</url>
      <title>DEV Community: Steven Hooley</title>
      <link>https://dev.to/sbhooley</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/sbhooley"/>
    <language>en</language>
    <item>
      <title>How we turned a flaky OpenClaw agent into a deterministic, 7.2 cheaper production workflow</title>
      <dc:creator>Steven Hooley</dc:creator>
      <pubDate>Sat, 11 Apr 2026 02:43:19 +0000</pubDate>
      <link>https://dev.to/sbhooley/how-we-turned-a-flaky-openclaw-agent-into-a-deterministic-72x-cheaper-production-workflow-40b4</link>
      <guid>https://dev.to/sbhooley/how-we-turned-a-flaky-openclaw-agent-into-a-deterministic-72x-cheaper-production-workflow-40b4</guid>
      <description>&lt;p&gt;&lt;strong&gt;How to Slash LLM Token and API Costs in OpenClaw Using AINL: Real-World Compile-Once Workflows That Actually Deliver 7.2× Savings&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;OpenClaw makes spinning up autonomous agents ridiculously easy—cron-triggered heartbeats, messaging integrations, tool calls, the works. But if your “simple” recurring flows (monitors, automations, data pulls, alerts) keep invoking an LLM for orchestration on every run, those costs compound fast. One production deployment later and your Anthropic, OpenRouter, or OpenAI bill shows a surprise top-line item.&lt;/p&gt;

&lt;p&gt;That’s exactly where we hit the wall. We rebuilt the same OpenClaw-style recurring jobs with &lt;strong&gt;AI Native Lang (AINL)&lt;/strong&gt;—a graph-based language that compiles your workflow once into deterministic, auditable code. The result? &lt;strong&gt;No runtime orchestration LLM calls&lt;/strong&gt; on steady-state execution, strict compile-time validation, and a measured &lt;strong&gt;7.2× cost reduction&lt;/strong&gt; across 17 live 24/7 cron jobs.&lt;/p&gt;

&lt;p&gt;This isn’t theory or marketing fluff. It’s live operational data from production infrastructure (as of March 2026). Here’s how anyone running OpenClaw can realistically apply the same pattern to cut token spend on &lt;strong&gt;monitoring, social automation, data processing, alert pipelines, report generation, and more&lt;/strong&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Core Problem: Why OpenClaw Agents Become Money Pits
&lt;/h3&gt;

&lt;p&gt;OpenClaw’s strength—flexible, LLM-driven agents with heartbeats, crons, and tool orchestration—becomes expensive at scale because most example patterns do this:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Every scheduled run wakes an agent that uses an LLM to “decide” the next step, interpret state, branch logic, and handle edge cases.&lt;/li&gt;
&lt;li&gt;Control flow lives in prompts instead of code.&lt;/li&gt;
&lt;li&gt;Costs scale linearly with frequency: 48 runs/day? 48 full orchestration calls.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Real examples from OpenClaw users (and our own fleet before the switch):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="http://ainativelang.com/blog/showcase-email-monitor-personal" rel="noopener noreferrer"&gt;Email/inbox monitors&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="http://ainativelang.com/blog/showcase-token-cost-tracker-enterprise" rel="noopener noreferrer"&gt;Metric threshold watchers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="http://ainativelang.com/blog/showcase-twitter-bot-x-promoter" rel="noopener noreferrer"&gt;Social media (X/Twitter) post classifiers or engagement scorers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="http://ainativelang.com/blog/showcase-morning-briefing-agent" rel="noopener noreferrer"&gt;Daily/Weekly report generators or data digest pipelines&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="http://ainativelang.com/blog/showcase-price-monitor-ecommerce" rel="noopener noreferrer"&gt;Track Prices or Financial Symbols&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even with optimizations like model tiering (cheap models for heartbeats) or LiteLLM caching, the orchestration layer still dominates recurring spend.&lt;/p&gt;

&lt;h3&gt;
  
  
  How AINL Flips the Economics: Compile Once, Run Deterministically
&lt;/h3&gt;

&lt;p&gt;AINL lets you author workflows as typed &lt;code&gt;.ainl&lt;/code&gt; graph files (slices, conditions, gates, adapters). The compiler turns it into a canonical graph IR with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Strict validation (catches schema mismatches, type errors, bad wiring &lt;strong&gt;before&lt;/strong&gt; deployment).&lt;/li&gt;
&lt;li&gt;Deterministic execution (same inputs → same path, every time).&lt;/li&gt;
&lt;li&gt;Adapter-based side effects (API calls, DB writes, queues) that are explicit and auditable via JSONL execution tapes.&lt;/li&gt;
&lt;li&gt;Zero runtime LLM calls for control flow in normal operation.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;LLM usage moves to &lt;strong&gt;authoring and compilation time only&lt;/strong&gt; (or intentional “reasoning” nodes you explicitly add). OpenClaw’s cron just calls the compiled &lt;code&gt;ainl-run&lt;/code&gt; binary or MCP-wired skill. No more per-tick orchestration prompts.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Live proof from the public AINL Cost Savings Report&lt;/strong&gt; (17 cron jobs, mix of monitoring + X automation):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Traditional agent-style loops: &lt;strong&gt;$7.00/day&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;AINL-compiled: &lt;strong&gt;$0.97/day&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7.2× reduction&lt;/strong&gt; (86% savings)&lt;/li&gt;
&lt;li&gt;Breakdown: X post generation (24/day), tweet classification (48/day), engagement scoring (48/day), plus 14 other intelligence/monitoring jobs.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Daily token burn drops because orchestration is paid once at compile time, not forever on every run.&lt;/p&gt;

&lt;h3&gt;
  
  
  Beyond Monitoring: Realistic Use Cases Where AINL Saves Real Money in OpenClaw
&lt;/h3&gt;

&lt;p&gt;The pattern shines on any &lt;strong&gt;recurring, policy-bound, or data-driven workflow&lt;/strong&gt; that doesn’t require fresh creative reasoning on every tick:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Monitoring &amp;amp; Alerting&lt;/strong&gt; (classic win)&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Inbox/email → filter + escalate/queue&lt;/li&gt;
&lt;li&gt;Metric/API polling → threshold gate → webhook/Slack/Telegram alert&lt;/li&gt;
&lt;li&gt;“Check DB or service health → notify only on anomaly”&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Social &amp;amp; Content Automation&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;X/Twitter search → classify → score engagement → auto-post or flag (the exact workflows in the cost report)&lt;/li&gt;
&lt;li&gt;Scheduled content digest or reply triage&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Data Pipelines &amp;amp; Reporting&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Pull from APIs/DBs → transform/filter → generate CSV/JSON report → email or push to queue&lt;/li&gt;
&lt;li&gt;Weekly metric aggregation or compliance checks&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Customer/Support Workflows&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Message triage → route to queue or auto-respond with templated logic&lt;/li&gt;
&lt;li&gt;Lead monitoring → score → notify sales&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Hybrid Flows (Best of Both Worlds)&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Keep a lightweight LLM reasoning node only where needed (e.g., “summarize this anomaly”).&lt;/li&gt;
&lt;li&gt;Everything else (gates, adapters, loops) stays deterministic and free at runtime.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;If your OpenClaw job is mostly “check → decide branch → act,” AINL crushes it. Pure creative or highly variable tasks (e.g., open-ended research) may still benefit from traditional agent loops—but even then, you can extract the deterministic parts into AINL for hybrid savings.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step-by-Step: Migrate Your Most Expensive OpenClaw Flow to AINL
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Install AINL&lt;/strong&gt; (Python 3.10+)
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;   pip &lt;span class="nb"&gt;install &lt;/span&gt;ainativelang          &lt;span class="c"&gt;# or 'ainativelang[mcp]' for OpenClaw extras&lt;/span&gt;
   ainl init my-workflow
   &lt;span class="nb"&gt;cd &lt;/span&gt;my-workflow
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Author the Workflow Once&lt;/strong&gt; (&lt;code&gt;main.ainl&lt;/code&gt;)&lt;br&gt;
Use compact syntax to define inputs, gates, adapters, and outputs. No Python glue required—describe the graph.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Validate &amp;amp; Compile&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;   ainl check main.ainl &lt;span class="nt"&gt;--strict&lt;/span&gt;     &lt;span class="c"&gt;# Catches bugs early&lt;/span&gt;
   &lt;span class="c"&gt;# Compile for your runtime&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Wire into OpenClaw&lt;/strong&gt; (one-command integration)
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;   ainl &lt;span class="nb"&gt;install &lt;/span&gt;openclaw &lt;span class="nt"&gt;--workspace&lt;/span&gt; ~/.openclaw/workspace
   &lt;span class="c"&gt;# This merges MCP servers, env vars, registers crons, and sets up ainl-run&lt;/span&gt;
   ainl cron add ./main.ainl &lt;span class="nt"&gt;--cron&lt;/span&gt; &lt;span class="s2"&gt;"0 * * * *"&lt;/span&gt;   &lt;span class="c"&gt;# e.g., hourly&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Deploy &amp;amp; Monitor&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;OpenClaw cron now executes the compiled graph directly.&lt;/li&gt;
&lt;li&gt;Check &lt;code&gt;ainl status&lt;/code&gt; for token estimates and cost avoided.&lt;/li&gt;
&lt;li&gt;Every run produces a JSONL tape for auditing/replay.&lt;/li&gt;
&lt;li&gt;Deploy changes: edit → recompile → git push (or &amp;lt;30s live).&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Drop-in example graphs (monitoring, metric threshold, HTTP poll → webhook) are in the public repo—copy, adapt, done.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why This Scales for Serious OpenClaw Users
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;10+ recurring jobs?&lt;/strong&gt; Costs were probably already painful. AINL makes them predictable.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Production reliability:&lt;/strong&gt; No prompt drift, no edge-case prompt failures, full audit trail.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Future-proof:&lt;/strong&gt; Same graph emits to ZeroClaw, Hermes Agent, or other runtimes via the toolchain.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Measurable ROI:&lt;/strong&gt; &lt;code&gt;ainl status&lt;/code&gt; shows estimated cost avoided. The report proves it compounds across a fleet.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Caveats for realism:&lt;/strong&gt; Savings are highest on orchestration-heavy flows (90–95% of the win in the report). If your job is 90%+ LLM reasoning every run, gains are smaller—but you can still extract the routing/adapter layer. Start with your most boring, highest-frequency, lowest-creativity job (the one quietly eating your budget).&lt;/p&gt;

&lt;h3&gt;
  
  
  Try It Today on Your Most Expensive Cron
&lt;/h3&gt;

&lt;p&gt;Pick the monitor, poller, classifier, or digest that shows up on your bill. Re-express it as one &lt;code&gt;.ainl&lt;/code&gt; graph, wire it via the OpenClaw installer, and watch the agent-style token burn disappear.&lt;/p&gt;




&lt;h2&gt;
  
  
  Links and Next Steps
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;AINL cost report (7.2× savings):&lt;/strong&gt; AINL_COST_SAVINGS_REPORT.md. &lt;a href="https://github.com/sbhooley/ainativelang/blob/main/AINL_COST_SAVINGS_REPORT.md" rel="noopener noreferrer"&gt;github&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;AINL + OpenClaw install guide:&lt;/strong&gt; Easy install for OpenClaw, ZeroClaw, and Hermes. &lt;a href="https://www.ainativelang.com/quickstart" rel="noopener noreferrer"&gt;ainativelang&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Star on GitHub: &lt;a href="https://github.com/sbhooley/ainativelang" rel="noopener noreferrer"&gt;AI Native Lang GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Join the community: &lt;a href="https://t.me/AINL_Portal" rel="noopener noreferrer"&gt;Telegram&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>programming</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Why Your AI Agents are Burning Cash (And How to Fix It in 3 Minutes)</title>
      <dc:creator>Steven Hooley</dc:creator>
      <pubDate>Fri, 10 Apr 2026 22:52:01 +0000</pubDate>
      <link>https://dev.to/sbhooley/why-your-ai-agents-are-burning-cash-and-how-to-fix-it-in-3-minutes-1j19</link>
      <guid>https://dev.to/sbhooley/why-your-ai-agents-are-burning-cash-and-how-to-fix-it-in-3-minutes-1j19</guid>
      <description>&lt;p&gt;The promise of AI agents was simple: set them loose, and they’ll handle the rest. But if you’ve actually tried to put an agent into production, you’ve likely hit a wall.&lt;/p&gt;

&lt;p&gt;Maybe it’s the unpredictable costs that spike every time your agent loops through a prompt. Maybe it’s the lack of reliability — where an agent that worked perfectly yesterday suddenly decides to hallucinate its own control flow today. Or maybe it’s the black-box nature of prompt-based orchestration that keeps your security team up at night.&lt;/p&gt;

&lt;p&gt;The reality is that most AI tools today are built for conversations, not for production infrastructure. They lack a reliable execution layer.&lt;/p&gt;

&lt;p&gt;That’s where AI Native Lang (AINL) comes in. It’s the “runtime-shaped hole” in the AI stack that we’ve all been waiting for.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem: The “Prompt Loop” Tax
&lt;/h2&gt;

&lt;p&gt;Traditional AI agents rely on “prompt loops” for orchestration. Every time the agent needs to decide what to do next, it calls the LLM. This leads to three major issues:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Compounding Costs:&lt;/strong&gt; You’re paying for the same orchestration tokens over and over again.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Non-Determinism:&lt;/strong&gt; LLMs are probabilistic. They can drift, fail silently, or ignore your instructions.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Latency:&lt;/strong&gt; Waiting for an LLM to “think” about every step slows down your workflows.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  The Solution: Compile Once, Run Forever
&lt;/h2&gt;

&lt;p&gt;AINL takes a different approach. Instead of asking the LLM to orchestrate every single run, use it to author the workflow once. AINL then compiles that workflow into a deterministic, auditable production worker.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;“Turn vague LLM conversations into deterministic, auditable production workers.”&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;By moving the orchestration logic into a compiled graph IR (Intermediate Representation), AINL ensures that your agent behaves like real infrastructure — not a fragile chatbot.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F562xn4811g3h3k16ywuv.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F562xn4811g3h3k16ywuv.webp" alt="Image description: " width="800" height="506"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Developers are Switching to AINL
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz9kd3id3ib71crs3kurc.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz9kd3id3ib71crs3kurc.webp" alt="Image description: AINL " width="800" height="506"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Deterministic by Design&lt;/strong&gt;&lt;br&gt;
In AINL, orchestration lives in the code, not the model. This means the same input produces the same result every time. It’s inspectable, diffable, and auditable.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Massive Cost Savings&lt;/strong&gt;&lt;br&gt;
Early adopters are reporting 2–5x lower recurring token spend on high-frequency workflows. By eliminating recurring orchestration calls, you can run monitoring-style workloads at near-zero cost.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Native MCP Integration&lt;/strong&gt;&lt;br&gt;
AINL is built for the modern AI IDE. With native Model Context Protocol (MCP) support, it fits perfectly into your existing development workflow.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Not Just for CLI: ArmaraOS&lt;/strong&gt;&lt;br&gt;
For those who prefer a UI, AINL powers ArmaraOS (&lt;a href="https://ainativelang.com" rel="noopener noreferrer"&gt;available on our website&lt;/a&gt;), a desktop app that puts a full AI agent dashboard on your computer. You can run agents, automate tasks, and stay in control — all without touching the command line.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Final Thoughts:&lt;/strong&gt; The Future of AI is Complicated&lt;br&gt;
We are moving away from the era of “throwing prompts at a wall” and toward an era of AI-native engineering. AINL provides the tools to build agents that are as reliable as the rest of your stack.&lt;/p&gt;

&lt;p&gt;Whether you’re a solo developer looking to cut costs or an enterprise team needing SOC 2-aligned audit trails, AINL is the control center you’ve been missing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ready to take control of your AI?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Visit the website: &lt;a href="https://ainativelang.com" rel="noopener noreferrer"&gt;ainativelang.com&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Developer’s Site: &lt;a href="https://stevenhooley.com" rel="noopener noreferrer"&gt;www.stevenhooley.com&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Star on GitHub: &lt;a href="https://github.com/sbhooley/ainativelang" rel="noopener noreferrer"&gt;AI Native Lang GitHub&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Join the community: &lt;a href="https://t.me/+tK_3BmA3Kb5hZTFk" rel="noopener noreferrer"&gt;Telegram&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Credit/Author: &lt;a href="https://medium.com/@web3jedi" rel="noopener noreferrer"&gt;Ai Jedi&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>webdev</category>
      <category>python</category>
    </item>
  </channel>
</rss>
