Search for "how to run LLMs locally" and count the Nvidia logos. CUDA this, CUDA that. If you own AMD hardware — and statistically, a lot of you do — the local AI ecosystem has treated you like a second-class citizen for years.
That just changed.
Lemonade is an open-source, AMD-backed local AI server that handles LLM chat, image generation, speech synthesis, and transcription — all from a single install, all running on your hardware, all private. It hit 216 points on Hacker News this week, and the discussion thread tells you everything about why AMD users are paying attention.
🍋 What Lemonade actually is: A 2MB native C++ service that auto-configures for your AMD GPU, NPU, or CPU. It exposes an OpenAI-compatible API at
localhost:13305, meaning any app that talks to OpenAI (VS Code Copilot, Open WebUI, n8n, Continue, hundreds more) works out of the box — pointed at your own machine instead of the cloud. Zero tokens billed. Zero data leaving your network.
Why This Matters Right Now
The local AI movement has been building momentum for two years. Ollama proved the concept. LM Studio made it pretty. But both share a dirty secret: AMD support is an afterthought. ROCm drivers are a maze. Getting llama.cpp to build with the right GPU target is a weekend project. Most users give up.
Lemonade's value proposition is brutally simple: one install, it detects your hardware, it works.
"If you have an AMD machine and want to run local models with minimal headache… it's really the easiest method. This runs on my NAS, handles my home assistant setup." — HN commenter
But it's not just ease of use. Lemonade is the only open-source OpenAI-compatible server that offers AMD Ryzen AI NPU acceleration. That's a hardware advantage Nvidia literally cannot match — there is no Nvidia NPU in your laptop.
The Architecture: NPU + GPU Hybrid Execution
On Ryzen AI 300/400 series chips (Strix Point, Strix Halo), Lemonade splits the workload:
Prompt processing (prefill) → Offloaded to the NPU, which has superior compute throughput for this specific task. This minimizes Time To First Token (TTFT).
Token generation (decode) → Handed to the integrated GPU (iGPU) or discrete GPU, which has better memory bandwidth for sequential token generation.
This hybrid approach is why a Ryzen AI laptop can feel snappier than raw token-per-second numbers would suggest.
Benchmarks: What Can You Actually Expect?
These are from AMD's own benchmarks on a Ryzen AI 9 HX 370 laptop (Radeon 890M, 32GB LPDDR5X-7500) running DeepSeek-R1-Distill-Llama-8B at INT4:
| Context Length | Time to First Token | Tokens/Second |
|---|---|---|
| 128 tokens | 0.94s | 20.7 tok/s |
| 256 tokens | 1.14s | 20.5 tok/s |
| 512 tokens | 1.65s | 20.0 tok/s |
| 1024 tokens | 2.68s | 19.2 tok/s |
| 2048 tokens | 5.01s | 17.6 tok/s |
Those are integrated graphics numbers. Not a $1,500 discrete GPU — a laptop chip.
📊 Community benchmarks from Strix Halo (128GB): GPT-OSS 120B at ~50 tok/s • Qwen3-Coder-Next at 43 tok/s (Q4) • Qwen3.5 35B-A3B at 55 tok/s (Q4) • Qwen3.5 27B at 11-12 tok/s (Q4, dense architecture). Yes — a 120B parameter model running at 50 tokens/second on a desktop APU with no discrete GPU.
Setup: From Zero to Running in Under 5 Minutes
Windows (Recommended)
# 1. Download the installer from GitHub releases
# https://github.com/lemonade-sdk/lemonade/releases/latest
# Run Lemonade_Server_Installer.exe
# 2. Select your models during installation
# The installer auto-detects your GPU/NPU
# 3. Launch from desktop shortcut — that's it.
# Server runs at http://localhost:13305
Linux (Ubuntu/Fedora)
# Ubuntu (snap)
sudo snap install lemonade-server
# Fedora (RPM)
sudo dnf install lemonade-server
# Start the server
lemonade run Gemma-3-4b-it-GGUF
macOS (Beta)
# Install via the official installer
# https://lemonade-server.ai/install_options.html#macos
lemonade run Gemma-3-4b-it-GGUF
Once running, pulling and switching models is dead simple:
# Browse available models
lemonade list
# Pull and run a model
lemonade pull Gemma-3-4b-it-GGUF
lemonade run Gemma-3-4b-it-GGUF
# Multi-modality
lemonade run SDXL-Turbo # Image gen
lemonade run kokoro-v1 # Speech synthesis
lemonade run Whisper-Large-v3-Turbo # Transcription
Connecting Apps: The OpenAI-Compatible Trick
Because Lemonade exposes an OpenAI-standard API, any app that supports custom OpenAI endpoints works immediately:
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:13305/api/v1",
api_key="lemonade" # required but unused
)
response = client.chat.completions.create(
model="Llama-3.2-1B-Instruct-Hybrid",
messages=[{"role": "user", "content": "Explain quantum computing"}]
)
print(response.choices[0].message.content)
That same endpoint works with VS Code Copilot, Open WebUI, Continue, n8n, and any OpenAI SDK in Python, Node.js, Go, Rust, C#, Java, Ruby, or PHP.
Lemonade vs. Ollama: The Honest Comparison
| Feature | Lemonade | Ollama |
|---|---|---|
| Primary focus | AMD optimization + multi-modality | Cross-platform model serving |
| GPU support | ROCm (AMD), Vulkan, Metal (beta) | CUDA (Nvidia), ROCm, Metal |
| NPU support | ✅ XDNA2 (Ryzen AI 300/400) | ❌ None |
| Modalities | Chat, Vision, Image Gen, TTS, STT | Chat, Vision |
| API compatibility | OpenAI, Ollama, Anthropic | Ollama, OpenAI (partial) |
| Multiple models | ✅ Simultaneously | One at a time |
| Mobile app | ✅ iOS + Android | ❌ |
| Binary size | ~2MB (server) | ~200MB |
Bottom line: If you're on AMD hardware, Lemonade is the better choice. If you need Nvidia CUDA support or the simplest possible cross-platform install, Ollama is still the safer bet.
One HN user ran a direct comparison on an M1 Max MacBook:
"Model: qwen3.59b. Ollama completed in about 1:44. Lemonade completed in about 1:14. So it seems faster in this very limited test."
The NPU Question: Is It Worth It?
What NPUs are good for:
- Low-power "always-on" inference for small models (1-4B parameters)
- Accelerating prompt processing (prefill) in hybrid mode
- Running AI tasks without touching your GPU
What NPUs are NOT good for (yet):
- Running large models (>10B parameters)
- Matching discrete GPU speeds for raw token generation
⚠️ NPU reality check: The NPU kernels used by Lemonade's FastFlowLM backend are proprietary (free for reasonable commercial use). The llama.cpp GPU path remains fully open. If you're on a Strix Halo with 128GB RAM, the GPU path is fast enough that NPU acceleration is a nice-to-have.
What's Coming Next
The Lemonade roadmap is active:
- MLX support — for better Apple Silicon performance
- vLLM support — for high-throughput serving
- Enhanced custom model support — easier GGUF/ONNX imports from Hugging Face
With Ubuntu 26.04 LTS adding native AMD NPU support and Lemonade 10.0 shipping Linux NPU support via FastFlowLM, Linux users are getting first-class treatment too.
The Bigger Picture
The llama.cpp creator Georgi Gerganov just joined Hugging Face — a consolidation event for the open-source local AI stack. Google's TurboQuant paper demonstrated KV cache compression to 3 bits, potentially slashing memory requirements. The infrastructure for running capable AI on consumer hardware is converging fast.
"I find it very frustrating to get LLMs, diffusion, etc. working fast on AMD. It's way too much work." — HN commenter, explaining exactly why Lemonade exists
Lemonade exists because that frustration is real, widespread, and fixable. If you've got AMD silicon, give it a shot. The install is a few minutes, the API is standard, and the models are free.
Links:
- Lemonade Server — Official site
- GitHub Repository — Source code + releases
- AMD Developer Article — Technical deep-dive
- Hacker News Discussion — Community reactions
Originally published on ComputeLeap

Top comments (0)