<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: SuryaKumaran Sivakumaran</title>
    <description>The latest articles on DEV Community by SuryaKumaran Sivakumaran (@suryakumaran).</description>
    <link>https://dev.to/suryakumaran</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3862249%2F5afdf05c-ee03-446e-a377-28a5b6cd26e2.png</url>
      <title>DEV Community: SuryaKumaran Sivakumaran</title>
      <link>https://dev.to/suryakumaran</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/suryakumaran"/>
    <language>en</language>
    <item>
      <title>Architecting Secure Local-First AI Agents with NemoClaw, Podman, and Ollama</title>
      <dc:creator>SuryaKumaran Sivakumaran</dc:creator>
      <pubDate>Sun, 05 Apr 2026 12:51:02 +0000</pubDate>
      <link>https://dev.to/suryakumaran/architecting-secure-local-first-ai-agents-with-nemoclaw-podman-and-ollama-cn9</link>
      <guid>https://dev.to/suryakumaran/architecting-secure-local-first-ai-agents-with-nemoclaw-podman-and-ollama-cn9</guid>
      <description>&lt;p&gt;The Shift to Local-First Agentic AI&lt;br&gt;
As we move toward more autonomous systems, the "Data Sovereignty vs. Capability" debate is intensifying. For many organizations and researchers, sending proprietary data or research logs to cloud-based LLMs is a non-starter.&lt;/p&gt;

&lt;p&gt;During my work on AetherMind (a research knowledge graph project), I set out to architect a "Zero-Trust" local environment for AI agents. The goal was simple but the execution was complex:&lt;/p&gt;

&lt;p&gt;Inference: High-performance local LLMs via Ollama.&lt;/p&gt;

&lt;p&gt;Security: Kernel-level sandboxing via NVIDIA NemoClaw.&lt;/p&gt;

&lt;p&gt;Hardware: Utilizing the full power of an MSI Vector 16 HX (RTX-powered) while maintaining a clean separation between Windows and WSL2.&lt;/p&gt;

&lt;p&gt;The Architectural Challenge: The Networking Moat&lt;br&gt;
The primary hurdle in this "Local-First" stack is the network boundary.&lt;/p&gt;

&lt;p&gt;Ollama typically runs on the Windows host to get direct, low-latency access to the GPU.&lt;/p&gt;

&lt;p&gt;NemoClaw (and its OpenShell runtime) operates within WSL2 to leverage Linux-native security features like Landlock and seccomp.&lt;/p&gt;

&lt;p&gt;By default, these two environments are isolated. When a NemoClaw agent attempts to call a local model, the connection is refused. Most systems then "fail-over" to cloud providers like Anthropic or OpenAI. If you haven't configured a cloud API key, the agent simply crashes with an Auth Error.&lt;/p&gt;

&lt;p&gt;The Solution: A Reliable Cross-Bridge Utility&lt;br&gt;
I developed a specialized utility to bridge this gap without compromising security. It ensures that the WSL2 sandbox can dynamically discover and authenticate with the Windows-hosted inference engine, regardless of IP shifts after reboots.&lt;/p&gt;

&lt;p&gt;I’ve open-sourced this as nemoclaw-podman-wsl2.&lt;/p&gt;

&lt;p&gt;Technical Highlights:&lt;br&gt;
Dynamic IP Discovery: Uses the WSL nameserver to resolve the Windows Host IP on-the-fly.&lt;/p&gt;

&lt;p&gt;Provider Force-Registration: Overrides the default "Anthropic-first" logic by injecting the local Ollama provider at the top of the auth-profiles.json stack.&lt;/p&gt;

&lt;p&gt;Podman-Friendly: Specifically tested for developers who prefer Podman over Docker Desktop for local container management.&lt;/p&gt;

&lt;p&gt;How to Implement the Architecture&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Prepare the Windows Host (Ollama)
Set Ollama to listen on all interfaces so WSL2 can reach it:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;PowerShell&lt;/p&gt;

&lt;h1&gt;
  
  
  Run in Admin PowerShell
&lt;/h1&gt;

&lt;p&gt;Restart Ollama and ensure your Windows Firewall allows inbound traffic on port 11434.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Deploy the Bridge (WSL2)
Clone the utility and run the verified setup script:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Bash&lt;br&gt;
git clone &lt;a href="https://github.com/suryakumaran2611/nemoclaw-podman-wsl2" rel="noopener noreferrer"&gt;https://github.com/suryakumaran2611/nemoclaw-podman-wsl2&lt;/a&gt;&lt;br&gt;
cd nemoclaw-podman-wsl2&lt;br&gt;
chmod +x setup_nemoclaw.sh&lt;br&gt;
./setup_nemoclaw.sh&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Verify the Secure Loop
Once onboarded, connect to your sandbox and launch the interactive terminal:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Bash&lt;br&gt;
nemoclaw main connect&lt;/p&gt;

&lt;h1&gt;
  
  
  Inside the sandbox:
&lt;/h1&gt;

&lt;p&gt;openclaw tui&lt;br&gt;
Resources &amp;amp; Documentation&lt;br&gt;
📖 Interactive Guide: &lt;a href="https://suryakumaran2611.github.io/nemoclaw-podman-wsl2/" rel="noopener noreferrer"&gt;https://suryakumaran2611.github.io/nemoclaw-podman-wsl2/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;💻 GitHub Repository: &lt;a href="https://github.com/suryakumaran2611/nemoclaw-podman-wsl2" rel="noopener noreferrer"&gt;https://github.com/suryakumaran2611/nemoclaw-podman-wsl2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Final Thoughts&lt;br&gt;
Architecting for local-first AI isn't just about privacy; it's about control. By securing the "hands" of the agent in a NemoClaw sandbox while keeping the "brain" in a high-performance Ollama instance, we create a robust framework for the next generation of autonomous tools.&lt;/p&gt;

&lt;p&gt;I'm looking forward to seeing how others adapt this bridge for their own specialized hardware setups. If you have questions or want to contribute to the logic, let’s connect on GitHub or in the comments!&lt;/p&gt;

&lt;h1&gt;
  
  
  LocalAI #AgenticAI #NVIDIA #Ollama #FullStackEngineering
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>nvidia</category>
      <category>ollamalocal</category>
      <category>architecture</category>
    </item>
  </channel>
</rss>
