This article was originally published on selfhosting.sh. Check us out for 770+ self-hosting guides.
Quick Recommendation
For most self-hosters running Proxmox VE, a used Dell OptiPlex 7050/7060 Micro with an Intel i5, 32 GB DDR4, and a 500 GB NVMe boot drive is the sweet spot. It costs $150–$200, draws 15–25W, and handles 5–10 VMs or 20+ containers without breaking a sweat. If you need more — GPU passthrough, ZFS with ECC, or 10+ VMs — step up to a used workstation or rack server.
What Proxmox Needs from Hardware
Proxmox VE is a Type 1 hypervisor based on Debian Linux. It runs directly on bare metal, so hardware compatibility matters more than with Docker-only setups.
CPU
- VT-x/VT-d required. Intel VT-x (or AMD-V) for virtualization, VT-d (or AMD-Vi) for PCI passthrough. Every modern CPU supports VT-x, but check your BIOS — it's sometimes disabled by default.
- Core count matters more than clock speed. Each VM gets dedicated vCPUs. Plan 1–2 cores per lightweight VM, 4+ for heavy workloads (databases, transcoding).
- Intel generally has better IOMMU grouping for PCI passthrough. AMD is fine for standard VM use.
RAM
- RAM is the primary bottleneck in Proxmox. Each VM needs its own allocation — you can't share like containers do.
- 32 GB minimum for a serious Proxmox host. 64 GB if you plan to run 10+ VMs.
- ECC recommended for ZFS. Not strictly required, but ZFS without ECC means a memory error can corrupt your pool silently. If you're storing important data on ZFS, use ECC.
Storage
- NVMe for boot and VM disks. Proxmox itself needs maybe 32 GB. VMs benefit enormously from NVMe speeds — random I/O matters.
- Separate boot and data drives. Don't run VMs off the same drive as Proxmox OS if you can avoid it.
- ZFS or LVM-thin for storage pools. ZFS gives snapshots, checksums, and compression. LVM-thin gives thin provisioning with less RAM overhead.
- SATA/SAS HDDs for bulk storage. NAS VMs, media libraries, backups — use spinning disks behind a ZFS mirror or RAID-Z.
Networking
- At least 1 Gbps. Ideally 2+ NICs for separating management traffic from VM traffic.
- Intel NICs preferred. Realtek works but Intel I210/I225/I226 have better driver support in Proxmox.
- 10 GbE is worth it if you're doing iSCSI, Ceph, or running storage-heavy VMs. See our 10GbE networking guide.
Recommended Builds
Budget Build — Under $200
Used Dell OptiPlex 7050 Micro
| Spec | Details |
|---|---|
| CPU | Intel Core i5-7500T (4C/4T, 2.7 GHz) |
| RAM | 32 GB DDR4 (2x16 GB) |
| Boot drive | 500 GB NVMe (add a 2242 or 2280 M.2) |
| Networking | 1x Gigabit Intel I219-LM |
| Power | 35W TDP, ~15W idle |
| Price | $120–$180 (used, eBay/refurb) |
What you can run: 5–8 lightweight VMs (Pi-hole, Home Assistant, Nextcloud, Vaultwarden) or 20+ LXC containers. No GPU passthrough — no discrete GPU slot.
Pros:
- Incredibly cheap for the performance
- Tiny form factor (fits on a shelf)
- Low power draw
- Reliable business-class hardware
Cons:
- Max 32 GB RAM (2 DIMM slots)
- No PCIe expansion (no GPU passthrough)
- Single NIC
Mid-Range Build — $300–$500
Used Dell OptiPlex 7080 Tower or HP EliteDesk 800 G6 Tower
| Spec | Details |
|---|---|
| CPU | Intel Core i7-10700 (8C/16T, 2.9 GHz) |
| RAM | 64 GB DDR4 (2x32 GB) |
| Boot drive | 500 GB NVMe |
| Data drive | 1 TB NVMe for VM storage |
| Networking | 1x Gigabit Intel + optional PCIe NIC |
| Power | 65W TDP, ~25W idle |
| Price | $300–$450 (used) |
What you can run: 10–15 VMs running concurrently. GPU passthrough with a low-profile card. Plex transcoding in a VM. TrueNAS VM with HBA passthrough if you add drives.
Pros:
- 8 cores / 16 threads is serious VM capacity
- 64 GB RAM handles many VMs
- PCIe slot for GPU or 10 GbE NIC
- iGPU available for Quick Sync transcoding passthrough
Cons:
- Tower form factor takes more space
- Single PSU (no redundancy)
High-End Build — $500–$1,000
Used Dell PowerEdge T340 or HP ProLiant ML350 Gen10
| Spec | Details |
|---|---|
| CPU | Intel Xeon E-2278G (8C/16T, 3.4 GHz) or Xeon Silver 4210 |
| RAM | 128 GB ECC DDR4 |
| Boot drive | 2x 500 GB NVMe (ZFS mirror) |
| Data storage | 4x 4 TB SATA in RAID-Z1 |
| Networking | 2x Gigabit Intel + optional 10 GbE |
| Power | ~80W idle |
| Price | $600–$1,000 (used) |
What you can run: Everything. 20+ VMs, Ceph cluster node, TrueNAS with ZFS, GPU passthrough for transcoding or AI workloads, multiple networks with VLANs.
Pros:
- ECC RAM (critical for ZFS)
- Hot-swap drive bays
- iLO/iDRAC for remote management
- Built for 24/7 operation
- Multiple PCIe slots
Cons:
- Louder (server fans)
- Higher power draw (~80–120W idle)
- Larger form factor
- Higher electricity cost (~$85–$125/year at $0.12/kWh)
DIY Build — Custom
For maximum flexibility, build your own.
| Component | Recommendation | Price |
|---|---|---|
| CPU | Intel Core i5-13500 (14C/20T) or AMD Ryzen 5 5600 | $150–$200 |
| Motherboard | ASRock B660M with 4 DIMM slots, Intel I226-V NIC | $100–$130 |
| RAM | 64 GB DDR4 ECC (if board supports) or non-ECC | $80–$120 |
| Boot drive | 500 GB NVMe (WD SN770 or Samsung 980) | $40–$50 |
| Case | Fractal Design Node 304 or Jonsbo N2 | $80–$100 |
| PSU | Corsair SF450 or Seasonic 450W 80+ Gold | $60–$80 |
| Total | $510–$680 |
Pros: Choose exactly what you need. Easy to upgrade. Better IOMMU grouping with consumer Intel boards for passthrough.
Cons: More work. No remote management (unless you add a BMC card). No hot-swap bays (unless your case supports it).
CPU Comparison for Proxmox
| CPU | Cores/Threads | TDP | Passmark (Multi) | Best For |
|---|---|---|---|---|
| Intel N100 | 4C/4T | 6W | ~5,500 | Lightweight — 3-5 containers, no VMs |
| Intel i5-7500T | 4C/4T | 35W | ~5,800 | Budget Proxmox — 5-8 VMs |
| Intel i5-10400 | 6C/12T | 65W | ~12,600 | Mid-range — 8-12 VMs |
| Intel i7-10700 | 8C/16T | 65W | ~16,000 | Solid all-rounder — 10-15 VMs |
| Intel i5-13500 | 14C/20T | 65W | ~28,000 | High-performance — 15-20+ VMs |
| Xeon E-2278G | 8C/16T | 80W | ~15,500 | ECC + iGPU — ZFS + transcoding |
| AMD Ryzen 5 5600 | 6C/12T | 65W | ~22,000 | Budget performance — great multi-thread |
RAM Sizing Guide
| Use Case | Minimum RAM | Recommended |
|---|---|---|
| 3-5 LXC containers only | 8 GB | 16 GB |
| 5-10 lightweight VMs | 16 GB | 32 GB |
| 10-15 mixed VMs | 32 GB | 64 GB |
| 15+ VMs or ZFS with ARC | 64 GB | 128 GB |
| Ceph node | 64 GB | 128 GB+ |
ZFS ARC note: ZFS uses RAM for its adaptive replacement cache (ARC). By default, it'll consume up to 50% of system RAM. You can limit it, but allocating at least 8-16 GB for ARC gives significantly better storage performance. Factor this into your RAM budget.
Storage Configuration Tips
Boot Drive
- 500 GB NVMe is plenty for Proxmox OS + ISO storage + container templates
- Mirror two NVMe drives (ZFS mirror) if you want boot drive redundancy
- Don't use USB drives for boot — they wear out fast under Proxmox's logging
VM Storage
- NVMe for performance-critical VMs (databases, Nextcloud, Gitea)
- SATA SSD for general VMs (Pi-hole, Home Assistant, monitoring)
- HDD for bulk storage VMs (media servers, backup targets)
ZFS vs LVM-Thin
| Feature | ZFS | LVM-Thin |
|---|---|---|
| Snapshots | Yes (instant, efficient) | Yes |
| Checksums | Yes (data integrity) | No |
| Compression | Yes (lz4 is nearly free) | No |
| RAM overhead | High (ARC cache) | Low |
| ECC recommended | Yes | Not critical |
| Complexity | Medium | Low |
Recommendation: Use ZFS if you have 32+ GB RAM and care about data integrity. Use LVM-thin if RAM is tight or you just want simple thin provisioning.
Power Consumption and Running Costs
| Build | Idle Power | Load Power | Annual Cost ($0.12/kWh) |
|---|---|---|---|
| OptiPlex Micro (budget) | 12–18W | 35–50W | $13–$19/year |
| OptiPlex Tower (mid) | 20–30W | 80–120W | $21–$32/year |
| PowerEdge T340 (high) | 60–90W | 200–350W | $63–$95/year |
| DIY build | 25–40W | 100–180W | $26–$42/year |
What Can You Run on Each Build?
Budget (4C/32GB)
- Pi-hole (LXC — 512 MB)
- Home Assistant (VM — 2 GB)
- Vaultwarden (LXC — 512 MB)
- Uptime Kuma (LXC — 512 MB)
- Nextcloud (VM — 4 GB)
- Headroom for 2-3 more lightweight services
Mid-Range (8C/64GB)
- Everything above, plus:
- Jellyfin with iGPU transcoding (VM — 4 GB)
- Gitea (VM — 2 GB)
- Grafana + Prometheus (VM — 4 GB)
- TrueNAS VM for NAS storage (VM — 8-16 GB)
- 5+ additional services
High-End (8C+/128GB)
- Full homelab: 15-20+ VMs running simultaneously
- Ceph storage cluster node
- Windows VM for testing
- GPU passthrough for Plex/Jellyfin transcoding
- Development environments
- Kubernetes cluster (k3s across multiple VMs)
FAQ
Can I run Proxmox on an Intel N100 mini PC?
Technically yes, but it's a poor fit. The N100 has only 4 cores and most mini PCs max out at 16 GB RAM. Proxmox VMs need dedicated RAM allocations, so you'll run out fast. Use Docker directly on an N100 instead — see our Intel N100 guide. If you insist on Proxmox, stick to LXC containers only.
Do I need ECC RAM for Proxmox?
Not strictly. Proxmox runs fine on consumer non-ECC RAM. But if you're using ZFS (which you should for data integrity), ECC is strongly recommended. A single bit flip in RAM can corrupt ZFS metadata, and without ECC there's no detection. For a homelab with replaceable data, non-ECC is acceptable. For important data, get ECC.
Should I use Proxmox or just run Docker directly?
Use Docker directly if you're running only containers and want simplicity. Use Proxmox if you need: VMs (Windows, TrueNAS, pfSense), PCI passthrough, network isolation between workloads, or high availability. Proxmox adds overhead — don't use it unless you need what it offers.
How much storage do I need?
Proxmox OS: 32 GB minimum. VM storage: 500 GB–2 TB NVMe depending on workloads. Bulk storage: as much as you need for media, backups, etc. Start with 500 GB NVMe + whatever HDDs you have, expand later.
Can I add a GPU later for transcoding passthrough?
Yes, if your system has a PCIe slot. Desktop towers and rack servers support this. Mini PCs and SFF systems generally don't. Plan for this upfront if transcoding matters to you.
Top comments (0)