I bought an RTX 5090. 32GB VRAM. The most powerful consumer GPU on the planet.
Then I tried to run ComfyUI on Windows. It broke immediately.
RuntimeError: sm_120 is not compatible
Three days later, I had a fully working solution. I packaged it and open-sourced it:
Here's the whole story.
Why RTX 50-series Breaks Everything
NVIDIA's Blackwell architecture (RTX 5090/5080/5070) uses a new compute capability code called sm_120. The problem? PyTorch's stable release doesn't include kernels for it.
This means:
-
pip install torchβ doesn't work on Blackwell - You need PyTorch nightly with CUDA 13.0 (cu130)
- But then xformers (the standard ComfyUI speed boost) forces PyTorch back to stable
- And custom nodes silently pull stable PyTorch through their dependencies
It's a dependency trap. Every fix creates a new problem.
The 5 Rules I Discovered
After 3 days of trial and error, I distilled everything into 5 rules. Break any one of them and your environment dies.
Rule 1: Use PyTorch nightly cu130 β stable doesn't have sm_120 kernels.
Rule 2: Never install xformers β it force-downgrades PyTorch to stable. This is the trap that got me twice.
Rule 3: Strip torch from every requirements.txt β custom nodes list torch as a dependency, and pip will happily replace your nightly build with stable.
Rule 4: Verify PyTorch after every custom node install β run python -c "import torch; print(torch.__version__)" and check that it still says cu130.
Rule 5: Clear proxy environment variables β system proxies block pip and git silently on Windows.
π‘ Pro Tip: Rule 2 was the hardest to figure out. xformers installs successfully, ComfyUI starts fine, and then crashes mid-inference with
sm_120 is not compatible. You don't even realize PyTorch was downgraded until you check the version.
What I Built
I automated all 5 rules into a one-click setup:
git clone https://github.com/hiroki-abe-58/ComfyUI-Win-Blackwell.git
cd ComfyUI-Win-Blackwell
# Double-click setup.bat β done in ~20 minutes
What setup.bat handles:
- Python 3.13 environment
- PyTorch nightly cu130 (not stable, not cu128)
- triton-windows + torch.compile (replaces xformers)
- ComfyUI core + custom dependencies (with torch stripped out)
- 28 verified custom nodes
- Post-install verification that cu130 is still intact
I also built companion tools:
-
verify_env.pyβ Blackwell-specific environment checker (sm_120, cu130, Triton, torch.compile) -
fix_windows_compat.pyβ Converts Linux workflow JSON paths to Windows format, replaces SageAttention with SDPA -
update.batβ Updates everything while preserving Blackwell compatibility
What I Verified
I tested 28 custom nodes one by one. Install β check PyTorch version β run test β record result. That was the most tedious part.
I also tested 5 Image-to-Video pipelines on 32GB VRAM:
- HunyuanVideo 1.5 I2V (8.3B params, ~16GB) β Smooth. My top recommendation.
- Kandinsky 5.0 Lite I2V (2B, ~4GB) β Very smooth. Great for quick tests.
- LTX-2 I2V (19B, ~25GB) β Works in FP8. Tight but fine.
- LongCat-Video TI2V (13.6B, ~14.5GB) β Works with adjustments.
- Kandinsky 5.0 Pro I2V (19B, ~40GB) β Needs CPU offload. Slow.
Why Not Just Use WSL2 or Docker?
The short answer: performance.
Loading safetensors through WSL2's NTFS translation layer is noticeably slower. Docker has the same issue plus additional setup complexity. For a tool like ComfyUI where you're iterating on workflows and loading large models frequently, native Windows file I/O makes a real difference.
Also, most AI artists using ComfyUI on Windows aren't Docker experts. A .bat file they can double-click is the right UX.
Wrapping Up
If you have an RTX 5090/5080/5070 and want to run ComfyUI on Windows without WSL2 or Docker, give it a try:
github.com/hiroki-abe-58/ComfyUI-Win-Blackwell
ComfyUI for GeForce RTX 50-Series (Blackwell)
The first fully documented, Windows-native ComfyUI setup for NVIDIA GeForce RTX 5090/5080/5070 (Blackwell architecture, sm_120) with CUDA 13.0.
What Makes This Special
RTX 50-series GPUs (Blackwell, Compute Capability sm_120) are not supported by PyTorch stable releases as of early 2026. Running ComfyUI on these GPUs requires specific versions and workarounds that are not documented anywhere else in a single, reproducible package.
Technical Highlights
Feature
Details
GPU Architecture
NVIDIA Blackwell (sm_120) -- RTX 5090 / 5080 / 5070
CUDA Version
13.0 (cu130) -- the latest CUDA runtime
PyTorch
Nightly cu130 build (not stable, not cu128)
Python
3.13 (latest)
Triton
triton-windows fork (official Triton is Linux-only)
xformers
Deliberately excluded (causes PyTorch downgrade)
Custom Nodes
28 verified nodes including video & music generation
Platform
Windows Native (no WSL2, no Docker required)
Why This Is Unique
-
Blackwell + Windows Native +β¦
MIT licensed. Stars and PRs welcome β especially if you verify additional custom nodes on Blackwell hardware.
Have you tried running AI tools on RTX 50-series? What was your experience? Let me know in the comments! π
If you found this helpful, consider following me for more AI + GPU content!
π Japanese version: Qiita / Zenn
π¦ Follow me on X: @geneLab_999
π» GitHub: hiroki-abe-58
Top comments (0)