<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Steven Aragón Urrea</title>
    <description>The latest articles on DEV Community by Steven Aragón Urrea (@stevearagonsite).</description>
    <link>https://dev.to/stevearagonsite</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1580013%2Ffc0a3b5a-bbdd-4dd9-8e66-1defcb6718d5.jpg</url>
      <title>DEV Community: Steven Aragón Urrea</title>
      <link>https://dev.to/stevearagonsite</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/stevearagonsite"/>
    <language>en</language>
    <item>
      <title>Experience Working with OpenClaw (Clawbot)</title>
      <dc:creator>Steven Aragón Urrea</dc:creator>
      <pubDate>Wed, 18 Mar 2026 18:58:24 +0000</pubDate>
      <link>https://dev.to/stevearagonsite/experience-working-with-openclaw-clawbot-298g</link>
      <guid>https://dev.to/stevearagonsite/experience-working-with-openclaw-clawbot-298g</guid>
      <description>&lt;h1&gt;
  
  
  Experience Working with OpenClaw (Clawbot)
&lt;/h1&gt;

&lt;p&gt;I want to share my experience as a developer working with OpenClaw (Clawbot), including a real-world setup and some practical insights after using it in production-like scenarios.&lt;/p&gt;

&lt;p&gt;My setup is based on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;AMD Ryzen 5 5600X&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;32GB RAM&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;RTX 3060 (LHR)&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;NVMe SSD storage&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Ubuntu Server (headless environment)&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I also experimented with multiple &lt;strong&gt;Ollama local models&lt;/strong&gt; as part of a fallback strategy, along with &lt;strong&gt;cloud models like Kimi 2.5&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;From a configuration standpoint, OpenClaw is a powerful and flexible system — but that flexibility comes at a cost.&lt;/p&gt;




&lt;h2&gt;
  
  
  Real Setup (Explained)
&lt;/h2&gt;

&lt;p&gt;Instead of showing raw configuration, here’s how my setup is structured conceptually:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Primary model:&lt;/strong&gt; Claude Opus 4.6
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cloud alternative:&lt;/strong&gt; Kimi 2.5
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Fallback #1:&lt;/strong&gt; Local models via Ollama (GPU)
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Fallback #2:&lt;/strong&gt; OpenAI models
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Interface:&lt;/strong&gt; Telegram
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Remote access:&lt;/strong&gt; Tailscale as a service
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Flow:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Primary cloud model → Alternative cloud → Local fallback → Secondary cloud fallback&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;This setup aims to balance:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Performance&lt;/li&gt;
&lt;li&gt;Reliability&lt;/li&gt;
&lt;li&gt;Offline capability&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Infrastructure &amp;amp; Tooling
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Ubuntu Server
&lt;/h3&gt;

&lt;p&gt;I chose &lt;strong&gt;Ubuntu Server&lt;/strong&gt; to fully utilize the machine:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Lower overhead (no GUI)&lt;/li&gt;
&lt;li&gt;Better resource allocation for models&lt;/li&gt;
&lt;li&gt;More predictable performance for long-running processes&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  Tailscale (as a service)
&lt;/h3&gt;

&lt;p&gt;I used &lt;strong&gt;Tailscale&lt;/strong&gt; running as a background service to access the machine remotely.&lt;/p&gt;

&lt;p&gt;The experience was &lt;strong&gt;excellent&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fast and stable connections&lt;/li&gt;
&lt;li&gt;Zero-config networking&lt;/li&gt;
&lt;li&gt;Secure remote access without exposing ports&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This made it extremely easy to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Manage OpenClaw remotely&lt;/li&gt;
&lt;li&gt;Debug issues&lt;/li&gt;
&lt;li&gt;Interact with the system from anywhere&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  Claude Code for Setup
&lt;/h3&gt;

&lt;p&gt;I used &lt;strong&gt;Claude Code&lt;/strong&gt; to bootstrap and configure the environment.&lt;/p&gt;

&lt;p&gt;This significantly reduced setup friction:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Faster iteration&lt;/li&gt;
&lt;li&gt;Easier debugging&lt;/li&gt;
&lt;li&gt;Better guidance wiring models and fallbacks&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Local Models Tested (Ollama)
&lt;/h2&gt;

&lt;p&gt;I tested several local models using Ollama:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Gemma 3 (12B)&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Qwen 3 (14B, abliterated)&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Qwen 3.5 (9B)&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Model Performance Ranking
&lt;/h2&gt;

&lt;h3&gt;
  
  
  🧠 Overall Ranking (Reasoning + Speed)
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Kimi 2.5 (cloud)&lt;/strong&gt; → Best overall performance
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Gemma 3 (local)&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Qwen 3 (local)&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Qwen 3.5 (local)&lt;/strong&gt;
&lt;/li&gt;
&lt;/ol&gt;




&lt;h3&gt;
  
  
  ☁️ Provider Ranking
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Anthropic (Claude models)&lt;/strong&gt; → Most reliable reasoning
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;OpenAI models&lt;/strong&gt; → Strong and consistent
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ollama (local models)&lt;/strong&gt; → Significantly weaker
&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  My Experience with Local Models
&lt;/h2&gt;

&lt;p&gt;Using &lt;strong&gt;Ollama with local models&lt;/strong&gt; is a great idea in theory.&lt;/p&gt;

&lt;h3&gt;
  
  
  Pros:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Works without internet&lt;/li&gt;
&lt;li&gt;Fully local&lt;/li&gt;
&lt;li&gt;Good fallback strategy&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Reality:
&lt;/h3&gt;

&lt;p&gt;Even running on an RTX 3060 and testing multiple models:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Local models were, in practice, a &lt;strong&gt;major downgrade&lt;/strong&gt;.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It almost feels like the system gets “lobotomized” when switching to them:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Weak reasoning&lt;/li&gt;
&lt;li&gt;Poor context handling&lt;/li&gt;
&lt;li&gt;Inconsistent outputs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This becomes very clear when compared to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Claude (Anthropic)&lt;/li&gt;
&lt;li&gt;OpenAI models&lt;/li&gt;
&lt;li&gt;Kimi 2.5 (which performed surprisingly well)&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Key Challenges
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Opacity in the TUI
&lt;/h3&gt;

&lt;p&gt;The TUI feels like a black box.&lt;/p&gt;

&lt;p&gt;You don’t know:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which model is active&lt;/li&gt;
&lt;li&gt;When fallbacks trigger&lt;/li&gt;
&lt;li&gt;Why decisions are made&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This makes debugging painful.&lt;/p&gt;




&lt;h3&gt;
  
  
  2. Lack of Cross-Channel Consistency
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Telegram ≠ TUI&lt;/li&gt;
&lt;li&gt;No shared continuity&lt;/li&gt;
&lt;li&gt;Fragmented sessions&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  3. Configuration Complexity
&lt;/h3&gt;

&lt;p&gt;You must carefully align:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Models&lt;/li&gt;
&lt;li&gt;Providers&lt;/li&gt;
&lt;li&gt;Fallbacks&lt;/li&gt;
&lt;li&gt;Channels&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Otherwise:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Silent failures&lt;/li&gt;
&lt;li&gt;Weird behaviors&lt;/li&gt;
&lt;li&gt;Hard-to-debug issues&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;OpenClaw has a &lt;strong&gt;strong architectural foundation&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Multi-model orchestration&lt;/li&gt;
&lt;li&gt;Fallback strategies&lt;/li&gt;
&lt;li&gt;Multi-channel interaction&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But it needs improvements in:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Transparency (TUI)&lt;/li&gt;
&lt;li&gt;Cross-channel consistency&lt;/li&gt;
&lt;li&gt;Developer experience&lt;/li&gt;
&lt;li&gt;Local model performance&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Combining:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Ubuntu Server&lt;/li&gt;
&lt;li&gt;Tailscale&lt;/li&gt;
&lt;li&gt;Cloud + local models&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;creates a &lt;strong&gt;very powerful personal AI infrastructure&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;However:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Today, &lt;strong&gt;cloud models still massively outperform local ones&lt;/strong&gt;, even on decent hardware like an RTX 3060.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The idea is solid.&lt;/p&gt;

&lt;p&gt;The execution is promising.&lt;/p&gt;

&lt;p&gt;But the ecosystem — especially around local models — still has a long way to go.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>linux</category>
      <category>llm</category>
      <category>openclaw</category>
    </item>
  </channel>
</rss>
