DEV Community

selfhosting.sh
selfhosting.sh

Posted on • Originally published at selfhosting.sh

Best Hardware for Proxmox VE in 2026

This article was originally published on selfhosting.sh. Check us out for 770+ self-hosting guides.


Quick Recommendation

For most self-hosters running Proxmox VE, a used Dell OptiPlex 7050/7060 Micro with an Intel i5, 32 GB DDR4, and a 500 GB NVMe boot drive is the sweet spot. It costs $150–$200, draws 15–25W, and handles 5–10 VMs or 20+ containers without breaking a sweat. If you need more — GPU passthrough, ZFS with ECC, or 10+ VMs — step up to a used workstation or rack server.

What Proxmox Needs from Hardware

Proxmox VE is a Type 1 hypervisor based on Debian Linux. It runs directly on bare metal, so hardware compatibility matters more than with Docker-only setups.

CPU

  • VT-x/VT-d required. Intel VT-x (or AMD-V) for virtualization, VT-d (or AMD-Vi) for PCI passthrough. Every modern CPU supports VT-x, but check your BIOS — it's sometimes disabled by default.
  • Core count matters more than clock speed. Each VM gets dedicated vCPUs. Plan 1–2 cores per lightweight VM, 4+ for heavy workloads (databases, transcoding).
  • Intel generally has better IOMMU grouping for PCI passthrough. AMD is fine for standard VM use.

RAM

  • RAM is the primary bottleneck in Proxmox. Each VM needs its own allocation — you can't share like containers do.
  • 32 GB minimum for a serious Proxmox host. 64 GB if you plan to run 10+ VMs.
  • ECC recommended for ZFS. Not strictly required, but ZFS without ECC means a memory error can corrupt your pool silently. If you're storing important data on ZFS, use ECC.

Storage

  • NVMe for boot and VM disks. Proxmox itself needs maybe 32 GB. VMs benefit enormously from NVMe speeds — random I/O matters.
  • Separate boot and data drives. Don't run VMs off the same drive as Proxmox OS if you can avoid it.
  • ZFS or LVM-thin for storage pools. ZFS gives snapshots, checksums, and compression. LVM-thin gives thin provisioning with less RAM overhead.
  • SATA/SAS HDDs for bulk storage. NAS VMs, media libraries, backups — use spinning disks behind a ZFS mirror or RAID-Z.

Networking

  • At least 1 Gbps. Ideally 2+ NICs for separating management traffic from VM traffic.
  • Intel NICs preferred. Realtek works but Intel I210/I225/I226 have better driver support in Proxmox.
  • 10 GbE is worth it if you're doing iSCSI, Ceph, or running storage-heavy VMs. See our 10GbE networking guide.

Recommended Builds

Budget Build — Under $200

Used Dell OptiPlex 7050 Micro

Spec Details
CPU Intel Core i5-7500T (4C/4T, 2.7 GHz)
RAM 32 GB DDR4 (2x16 GB)
Boot drive 500 GB NVMe (add a 2242 or 2280 M.2)
Networking 1x Gigabit Intel I219-LM
Power 35W TDP, ~15W idle
Price $120–$180 (used, eBay/refurb)

What you can run: 5–8 lightweight VMs (Pi-hole, Home Assistant, Nextcloud, Vaultwarden) or 20+ LXC containers. No GPU passthrough — no discrete GPU slot.

Pros:

  • Incredibly cheap for the performance
  • Tiny form factor (fits on a shelf)
  • Low power draw
  • Reliable business-class hardware

Cons:

  • Max 32 GB RAM (2 DIMM slots)
  • No PCIe expansion (no GPU passthrough)
  • Single NIC

Mid-Range Build — $300–$500

Used Dell OptiPlex 7080 Tower or HP EliteDesk 800 G6 Tower

Spec Details
CPU Intel Core i7-10700 (8C/16T, 2.9 GHz)
RAM 64 GB DDR4 (2x32 GB)
Boot drive 500 GB NVMe
Data drive 1 TB NVMe for VM storage
Networking 1x Gigabit Intel + optional PCIe NIC
Power 65W TDP, ~25W idle
Price $300–$450 (used)

What you can run: 10–15 VMs running concurrently. GPU passthrough with a low-profile card. Plex transcoding in a VM. TrueNAS VM with HBA passthrough if you add drives.

Pros:

  • 8 cores / 16 threads is serious VM capacity
  • 64 GB RAM handles many VMs
  • PCIe slot for GPU or 10 GbE NIC
  • iGPU available for Quick Sync transcoding passthrough

Cons:

  • Tower form factor takes more space
  • Single PSU (no redundancy)

High-End Build — $500–$1,000

Used Dell PowerEdge T340 or HP ProLiant ML350 Gen10

Spec Details
CPU Intel Xeon E-2278G (8C/16T, 3.4 GHz) or Xeon Silver 4210
RAM 128 GB ECC DDR4
Boot drive 2x 500 GB NVMe (ZFS mirror)
Data storage 4x 4 TB SATA in RAID-Z1
Networking 2x Gigabit Intel + optional 10 GbE
Power ~80W idle
Price $600–$1,000 (used)

What you can run: Everything. 20+ VMs, Ceph cluster node, TrueNAS with ZFS, GPU passthrough for transcoding or AI workloads, multiple networks with VLANs.

Pros:

  • ECC RAM (critical for ZFS)
  • Hot-swap drive bays
  • iLO/iDRAC for remote management
  • Built for 24/7 operation
  • Multiple PCIe slots

Cons:

  • Louder (server fans)
  • Higher power draw (~80–120W idle)
  • Larger form factor
  • Higher electricity cost (~$85–$125/year at $0.12/kWh)

DIY Build — Custom

For maximum flexibility, build your own.

Component Recommendation Price
CPU Intel Core i5-13500 (14C/20T) or AMD Ryzen 5 5600 $150–$200
Motherboard ASRock B660M with 4 DIMM slots, Intel I226-V NIC $100–$130
RAM 64 GB DDR4 ECC (if board supports) or non-ECC $80–$120
Boot drive 500 GB NVMe (WD SN770 or Samsung 980) $40–$50
Case Fractal Design Node 304 or Jonsbo N2 $80–$100
PSU Corsair SF450 or Seasonic 450W 80+ Gold $60–$80
Total $510–$680

Pros: Choose exactly what you need. Easy to upgrade. Better IOMMU grouping with consumer Intel boards for passthrough.

Cons: More work. No remote management (unless you add a BMC card). No hot-swap bays (unless your case supports it).

CPU Comparison for Proxmox

CPU Cores/Threads TDP Passmark (Multi) Best For
Intel N100 4C/4T 6W ~5,500 Lightweight — 3-5 containers, no VMs
Intel i5-7500T 4C/4T 35W ~5,800 Budget Proxmox — 5-8 VMs
Intel i5-10400 6C/12T 65W ~12,600 Mid-range — 8-12 VMs
Intel i7-10700 8C/16T 65W ~16,000 Solid all-rounder — 10-15 VMs
Intel i5-13500 14C/20T 65W ~28,000 High-performance — 15-20+ VMs
Xeon E-2278G 8C/16T 80W ~15,500 ECC + iGPU — ZFS + transcoding
AMD Ryzen 5 5600 6C/12T 65W ~22,000 Budget performance — great multi-thread

RAM Sizing Guide

Use Case Minimum RAM Recommended
3-5 LXC containers only 8 GB 16 GB
5-10 lightweight VMs 16 GB 32 GB
10-15 mixed VMs 32 GB 64 GB
15+ VMs or ZFS with ARC 64 GB 128 GB
Ceph node 64 GB 128 GB+

ZFS ARC note: ZFS uses RAM for its adaptive replacement cache (ARC). By default, it'll consume up to 50% of system RAM. You can limit it, but allocating at least 8-16 GB for ARC gives significantly better storage performance. Factor this into your RAM budget.

Storage Configuration Tips

Boot Drive

  • 500 GB NVMe is plenty for Proxmox OS + ISO storage + container templates
  • Mirror two NVMe drives (ZFS mirror) if you want boot drive redundancy
  • Don't use USB drives for boot — they wear out fast under Proxmox's logging

VM Storage

  • NVMe for performance-critical VMs (databases, Nextcloud, Gitea)
  • SATA SSD for general VMs (Pi-hole, Home Assistant, monitoring)
  • HDD for bulk storage VMs (media servers, backup targets)

ZFS vs LVM-Thin

Feature ZFS LVM-Thin
Snapshots Yes (instant, efficient) Yes
Checksums Yes (data integrity) No
Compression Yes (lz4 is nearly free) No
RAM overhead High (ARC cache) Low
ECC recommended Yes Not critical
Complexity Medium Low

Recommendation: Use ZFS if you have 32+ GB RAM and care about data integrity. Use LVM-thin if RAM is tight or you just want simple thin provisioning.

Power Consumption and Running Costs

Build Idle Power Load Power Annual Cost ($0.12/kWh)
OptiPlex Micro (budget) 12–18W 35–50W $13–$19/year
OptiPlex Tower (mid) 20–30W 80–120W $21–$32/year
PowerEdge T340 (high) 60–90W 200–350W $63–$95/year
DIY build 25–40W 100–180W $26–$42/year

What Can You Run on Each Build?

Budget (4C/32GB)

Mid-Range (8C/64GB)

  • Everything above, plus:
  • Jellyfin with iGPU transcoding (VM — 4 GB)
  • Gitea (VM — 2 GB)
  • Grafana + Prometheus (VM — 4 GB)
  • TrueNAS VM for NAS storage (VM — 8-16 GB)
  • 5+ additional services

High-End (8C+/128GB)

  • Full homelab: 15-20+ VMs running simultaneously
  • Ceph storage cluster node
  • Windows VM for testing
  • GPU passthrough for Plex/Jellyfin transcoding
  • Development environments
  • Kubernetes cluster (k3s across multiple VMs)

FAQ

Can I run Proxmox on an Intel N100 mini PC?

Technically yes, but it's a poor fit. The N100 has only 4 cores and most mini PCs max out at 16 GB RAM. Proxmox VMs need dedicated RAM allocations, so you'll run out fast. Use Docker directly on an N100 instead — see our Intel N100 guide. If you insist on Proxmox, stick to LXC containers only.

Do I need ECC RAM for Proxmox?

Not strictly. Proxmox runs fine on consumer non-ECC RAM. But if you're using ZFS (which you should for data integrity), ECC is strongly recommended. A single bit flip in RAM can corrupt ZFS metadata, and without ECC there's no detection. For a homelab with replaceable data, non-ECC is acceptable. For important data, get ECC.

Should I use Proxmox or just run Docker directly?

Use Docker directly if you're running only containers and want simplicity. Use Proxmox if you need: VMs (Windows, TrueNAS, pfSense), PCI passthrough, network isolation between workloads, or high availability. Proxmox adds overhead — don't use it unless you need what it offers.

How much storage do I need?

Proxmox OS: 32 GB minimum. VM storage: 500 GB–2 TB NVMe depending on workloads. Bulk storage: as much as you need for media, backups, etc. Start with 500 GB NVMe + whatever HDDs you have, expand later.

Can I add a GPU later for transcoding passthrough?

Yes, if your system has a PCIe slot. Desktop towers and rack servers support this. Mini PCs and SFF systems generally don't. Plan for this upfront if transcoding matters to you.

Related

Top comments (0)