DEV Community

brian austin
brian austin

Posted on

sllm vs ANTHROPIC_BASE_URL: two ways to get unlimited Claude tokens

sllm vs ANTHROPIC_BASE_URL: two ways to get unlimited Claude tokens

Two approaches just landed on Hacker News in the same week: sllm (split a GPU node with other developers) and ANTHROPIC_BASE_URL proxies. Both promise unlimited tokens. They work very differently.

Here's the practical comparison.

The core difference

sllm gives you shared compute — you're literally splitting GPU time with strangers on a node. Cheap, but your inference speed depends on who else is using the node right now.

ANTHROPIC_BASE_URL proxies give you a managed Claude API endpoint. You set one environment variable, and your Claude Code sessions route through it. You still get Claude's actual models, not a local alternative.

# The entire setup for ANTHROPIC_BASE_URL approach:
export ANTHROPIC_BASE_URL=https://simplylouie.com/api
export ANTHROPIC_API_KEY=your-key
claude  # now using proxy, rate limits lifted
Enter fullscreen mode Exit fullscreen mode

That's it. One variable.

Side-by-side comparison

Feature sllm ANTHROPIC_BASE_URL proxy
Model Local/open weights Real Claude (Sonnet/Haiku)
Setup Run local server One env var
Rate limits None (your GPU slice) Lifted via proxy
Anthropic ToS N/A Compliant
CVE risk Depends on impl None (hosted)
Price Varies ✌️$2/month
Works with Claude Code Needs config Native
Context window Model-dependent Claude native

When to use sllm

sllm makes sense when:

  • You specifically want open weights (Llama, Mistral, etc.)
  • You're doing batch inference and don't need Claude specifically
  • You want to experiment with local models at scale
  • You're comfortable managing shared infrastructure

When to use ANTHROPIC_BASE_URL

ANTHROPIC_BASE_URL proxy makes sense when:

  • You're using Claude Code and hitting rate limits mid-session
  • You need Claude specifically (not a local alternative)
  • You want zero setup — one env var, done
  • You're in a team where everyone needs the same setup

The Claude Code use case

This is where ANTHROPIC_BASE_URL wins completely. Claude Code expects Anthropic's API. The ANTHROPIC_BASE_URL environment variable exists specifically to redirect those calls:

# In your shell profile:
export ANTHROPIC_BASE_URL=https://simplylouie.com/api
export ANTHROPIC_API_KEY=sl-your-key-here

# Now run Claude Code normally:
claude "refactor this auth module"

# No rate limits. Same Claude. Session continues uninterrupted.
Enter fullscreen mode Exit fullscreen mode

With sllm, you'd need to reconfigure Claude Code to point to a different model endpoint — and you'd lose Claude's specific capabilities.

The rate limit problem, specifically

Claude Code's rate limit hits mid-session. You're in the middle of a refactor. Claude stops:

Claude is currently unavailable. Please wait before continuing.
Enter fullscreen mode Exit fullscreen mode

The ANTHROPIC_BASE_URL approach means that message never appears. The proxy manages the rate limiting transparently — you keep coding.

Real-world setup (30 seconds)

# 1. Sign up at simplylouie.com
# 2. Get your API key from dashboard
# 3. Add to .bashrc or .zshrc:
export ANTHROPIC_BASE_URL=https://simplylouie.com/api
export ANTHROPIC_API_KEY=sl-your-key

# 4. Source it:
source ~/.bashrc

# 5. Run Claude Code:
claude
# Done. Rate limits gone.
Enter fullscreen mode Exit fullscreen mode

For CI/CD:

# .github/workflows/ai-review.yml
env:
  ANTHROPIC_BASE_URL: https://simplylouie.com/api
  ANTHROPIC_API_KEY: ${{ secrets.LOUIE_API_KEY }}
Enter fullscreen mode Exit fullscreen mode

The cost comparison

  • sllm: Variable (depends on GPU time purchased)
  • Claude Pro: $20/month (with rate limits)
  • SimplyLouie proxy: ✌️$2/month (rate limits lifted, real Claude)

For developers who specifically use Claude Code and just want the rate limits gone, the math is straightforward.

Bottom line

If you want open weight models at scale, sllm is interesting infrastructure.

If you want Claude Code to work without interruption, ANTHROPIC_BASE_URL is the correct tool. One env var. $2/month. Rate limits gone.

SimplyLouie ANTHROPIC_BASE_URL setup — 7-day free trial, no card required


Both approaches have their place. The question is whether you need Claude specifically or just need cheap inference tokens.

Top comments (0)