Most people think you need expensive APIs or a powerful GPU to run large AI models.
You don't. Here's how I ran a 397B parameter model for free in under 5 minutes on Windows.
What You Actually Need
- A Windows machine
- An Ollama account (free)
- Internet connection
That's it. No GPU. No API key. No Anthropic billing.
What's happening under the hood
Two things working together:
- Claude Code CLI is just the terminal interface and agent shell. Nothing goes through Anthropic's servers.
- Ollama hosts and runs Qwen3.5 397B on their own cloud infrastructure, completely free
Claude Code supports Ollama as a backend. So you get a familiar, powerful agent interface while Ollama handles 100% of the actual inference. Anthropic is not involved beyond providing the CLI shell.
Step 1: Install Claude Code
Open PowerShell and run:
irm https://claude.ai/install.ps1 | iex
Step 2: Install Ollama
irm https://ollama.com/install.ps1 | iex
Step 3: Launch Claude Code with the 397B model
ollama launch claude --model qwen3.5:397b-cloud
You'll be asked to verify with your Ollama account once. After that it just works.
What you get
- Full Claude Code agent interface
- 397B parameter model (Qwen3.5)
- ~256K token context window
- Cloud hosted, so no local GPU needed
- Free as long as you have an Ollama account
Important notes
- This routes through Ollama's cloud, so you need internet
- Don't use it for sensitive or private data
- Free access may change in the future, use it while it lasts
Why this matters
We're entering a phase where the agent layer and the model layer are completely separate.
Claude Code is just the interface. The model underneath can be swapped. Local or cloud, open or closed, free or paid.
This setup is a simple example of that shift. The barrier to running frontier-scale models is no longer hardware or money. It's just knowing how to connect the right tools.
Try it
Three commands. Less than 5 minutes. No credit card.
If you run into issues drop them in the comments.
I research LLM training and continual learning. Follow for more no-fluff AI content.
Top comments (0)