<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: zhing2006</title>
    <description>The latest articles on DEV Community by zhing2006 (@zhing2006).</description>
    <link>https://dev.to/zhing2006</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3842926%2Fe732a236-a856-4321-b455-e80f4c3b5b37.jpg</url>
      <title>DEV Community: zhing2006</title>
      <link>https://dev.to/zhing2006</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/zhing2006"/>
    <language>en</language>
    <item>
      <title>krew-cli lets you run GPT, Claude, Gemini (and any OpenAI-compatible provider) in a single terminal session</title>
      <dc:creator>zhing2006</dc:creator>
      <pubDate>Wed, 25 Mar 2026 09:56:49 +0000</pubDate>
      <link>https://dev.to/zhing2006/krew-cli-lets-you-run-gpt-claude-gemini-and-any-openai-compatible-provider-in-a-single-terminal-2klm</link>
      <guid>https://dev.to/zhing2006/krew-cli-lets-you-run-gpt-claude-gemini-and-any-openai-compatible-provider-in-a-single-terminal-2klm</guid>
      <description>&lt;p&gt;I built a CLI that lets you chat with multiple LLM models (GPT, Claude, Gemini) in one terminal.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frrtkzzr9zcvupnwrxiqi.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frrtkzzr9zcvupnwrxiqi.gif" alt=" " width="600" height="362"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;What if you could host a roundtable meeting with multiple AI models — all in your terminal?&lt;/p&gt;

&lt;p&gt;That's exactly what &lt;strong&gt;krew&lt;/strong&gt; does. It lets you run GPT, Claude, Gemini (and any OpenAI-compatible provider) in a single terminal session, with shared context so agents can see and build on each other's answers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Quick Start
&lt;/h2&gt;

&lt;p&gt;Get started in 3 commands:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;npm &lt;span class="nb"&gt;install&lt;/span&gt; &lt;span class="nt"&gt;-g&lt;/span&gt; @zhing2026/krew
krew config init
krew
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The config wizard walks you through setting up providers and agents — no manual config file editing needed.&lt;/p&gt;

&lt;h2&gt;
  
  
  @ Addressing — Talk to Any Agent
&lt;/h2&gt;

&lt;p&gt;Use &lt;code&gt;@&lt;/code&gt; to control who you're talking to:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;› @all What's the best way to handle errors in Rust?
  (all agents respond in order)

› @opus Can you elaborate on the Result type?
  (only Claude responds)

› Tell me more
  (continues with the last respondent)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  #Whisper — Private Messages
&lt;/h2&gt;

&lt;p&gt;Want to ask one agent to privately evaluate another's answer?&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;› @all Propose an architecture for a chat app
  (both agents answer publicly)

› #opus What are the weaknesses in GPT's proposal?
  (only opus sees this — other agents see a placeholder)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You can even create private whisper groups:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;› #opus #gemini Discuss the tradeoffs between these approaches
  (only opus and gemini see each other's replies)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  AI-to-AI Routing
&lt;/h2&gt;

&lt;p&gt;When an agent's reply &lt;code&gt;@mentions&lt;/code&gt; another agent, that agent is automatically dispatched. You can sit back and watch them collaborate (or argue).&lt;/p&gt;

&lt;h2&gt;
  
  
  Built-in Tools
&lt;/h2&gt;

&lt;p&gt;Agents aren't just chatbots — they can take action:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;Description&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;read_file&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Read file content&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;write_file&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Create or overwrite files&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;edit_file&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Search-and-replace editing&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;shell&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Execute shell commands&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;glob&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;File pattern matching&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;grep&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Content search with regex&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;fetch_url&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Fetch and parse web pages&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;activate_skill&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Load specialized skill instructions&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;All file operations are sandboxed to your project directory.&lt;/p&gt;

&lt;h2&gt;
  
  
  MCP Integration
&lt;/h2&gt;

&lt;p&gt;Extend agent capabilities via &lt;a href="https://modelcontextprotocol.io" rel="noopener noreferrer"&gt;Model Context Protocol&lt;/a&gt; servers — both stdio and HTTP transports are supported:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight toml"&gt;&lt;code&gt;&lt;span class="nn"&gt;[[mcp_servers]]&lt;/span&gt;
&lt;span class="py"&gt;name&lt;/span&gt; &lt;span class="p"&gt;=&lt;/span&gt; &lt;span class="s"&gt;"filesystem"&lt;/span&gt;
&lt;span class="py"&gt;command&lt;/span&gt; &lt;span class="p"&gt;=&lt;/span&gt; &lt;span class="s"&gt;"npx"&lt;/span&gt;
&lt;span class="py"&gt;args&lt;/span&gt; &lt;span class="p"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="s"&gt;"-y"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;"@modelcontextprotocol/server-filesystem"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;"."&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Skill System
&lt;/h2&gt;

&lt;p&gt;Define reusable skill packages with a &lt;code&gt;SKILL.md&lt;/code&gt; file. Agents automatically discover available skills and activate them when needed:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;my-skill/
├── SKILL.md          # Skill definition (name, description, instructions)
├── scripts/          # Helper scripts
└── references/       # Reference materials
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Custom Slash Commands
&lt;/h2&gt;

&lt;p&gt;Create your own commands as Markdown files with argument substitution and bash preprocessing:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="nn"&gt;---&lt;/span&gt;
&lt;span class="na"&gt;description&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;Review code for issues&lt;/span&gt;
&lt;span class="na"&gt;argument-hint&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;&amp;lt;file_path&amp;gt;&lt;/span&gt;
&lt;span class="nn"&gt;---&lt;/span&gt;

Please review the following file: $ARGUMENTS

Here are the recent changes:
!&lt;span class="sb"&gt;`git diff --cached`&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Save as &lt;code&gt;.krew/commands/review.md&lt;/code&gt;, then use &lt;code&gt;/review src/main.rs&lt;/code&gt; in your session.&lt;/p&gt;

&lt;h2&gt;
  
  
  Session Management
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Persistence&lt;/strong&gt; — Every message is saved in real-time. Crash? No problem.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Resume&lt;/strong&gt; — Pick up any previous conversation with &lt;code&gt;/resume&lt;/code&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Rewind&lt;/strong&gt; — Fork from any point in history with &lt;code&gt;/rewind&lt;/code&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Auto-compact&lt;/strong&gt; — Automatic context compression when token limits approach&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Prompt Mode for CI/CD
&lt;/h2&gt;

&lt;p&gt;Run one-shot prompts from scripts:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="c"&gt;# Code review in CI&lt;/span&gt;
git diff HEAD~1 | krew &lt;span class="nt"&gt;-p&lt;/span&gt; &lt;span class="s2"&gt;"@opus review these changes for bugs"&lt;/span&gt;

&lt;span class="c"&gt;# JSON output for parsing&lt;/span&gt;
krew &lt;span class="nt"&gt;-p&lt;/span&gt; &lt;span class="s2"&gt;"@all hello"&lt;/span&gt; &lt;span class="nt"&gt;--format&lt;/span&gt; json
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  More Features
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Streaming output&lt;/strong&gt; — Markdown rendering with syntax highlighting and per-agent color coding&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Thinking/reasoning&lt;/strong&gt; — Display model thinking process (configurable: low/medium/high)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Web search&lt;/strong&gt; — Provider-native web search (OpenAI, Anthropic, Gemini)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Per-agent sampling&lt;/strong&gt; — Configure temperature, top_p, max_tokens per agent&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Project instructions&lt;/strong&gt; — &lt;code&gt;AGENTS.md&lt;/code&gt; files auto-injected into system prompts&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Config management&lt;/strong&gt; — &lt;code&gt;krew config init/add/del/list/doctor&lt;/code&gt; for full config CRUD&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Supported Providers
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Provider&lt;/th&gt;
&lt;th&gt;Examples&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;OpenAI&lt;/td&gt;
&lt;td&gt;GPT-5.2&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Anthropic&lt;/td&gt;
&lt;td&gt;Claude Opus 4.6, Sonnet 4.6&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Google&lt;/td&gt;
&lt;td&gt;Gemini 3.1 Pro (+ Vertex AI)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;OpenAI-Compatible&lt;/td&gt;
&lt;td&gt;OpenRouter, LiteLLM, any compatible API&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  Built with Rust
&lt;/h2&gt;

&lt;p&gt;Single static binary. Zero runtime dependencies. 5 platforms (Windows, Linux x64/arm64, macOS x64/arm64).&lt;/p&gt;




&lt;p&gt;GitHub: &lt;a href="https://github.com/ZHing2006/krew-cli" rel="noopener noreferrer"&gt;https://github.com/ZHing2006/krew-cli&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Install: &lt;code&gt;npm install -g @zhing2026/krew&lt;/code&gt;&lt;br&gt;
Config: &lt;code&gt;krew config init&lt;/code&gt;&lt;br&gt;
Run: &lt;code&gt;krew&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Feedback and contributions welcome!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>cli</category>
      <category>rust</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
