Overview
I’m currently building a multi-purpose home lab server designed to run several workloads simultaneously, with the goal of consolidating media, virtualization, storage, local AI, and home automation into a single always-on platform.
Planned workloads:
• Plex Media Server
• Virtual machines and cybersecurity labs
• Network File Server (NAS)
• Local AI stack using OpenClaw paired with Llama models
• Home automation and camera integration
The idea is to have one system act as a media hub, development environment, and private AI infrastructure without relying on cloud services.
I’d like feedback on the build overall, but especially GPU recommendations toward the end.
Design Goals
The system is being built around a few priorities:
- Run multiple services concurrently without bottlenecks
- Support local LLM inference
- Reliable Plex streaming and hardware transcoding
- Dedicated environment for cybersecurity labs and testing
- Easy future GPU expansion
- Enterprise-level stability for 24/7 uptime
Rather than maintaining multiple smaller machines, I decided to consolidate everything onto a high-core enterprise platform.
Hardware Specifications
System Platform
Fully Built & Tested (ASRock Platform)
CPU
AMD EPYC 7452
32 Cores / 64 Threads
Chosen primarily for virtualization density and parallel workloads. The high core count allows containers, VMs, and AI services to run concurrently without stepping on each other.
Primary Responsibilities
- VM hosting
- Container orchestration
- AI services
- Background media processing
Memory
128 GB RAM (2 × 64 GB DIMMs)
Large RAM capacity is intended to support both virtualization and local LLM workloads.
Planned Allocation
- AI workloads: 32–64 GB
- Virtual machines: 32–48 GB
- Plex + system services: 8–12 GB
- Remaining memory reserved for caching
Power Supply
Corsair HX1500i (1500W Platinum)
Oversized intentionally to allow future GPU expansion without power constraints.
High-Speed Storage
Samsung 990 Pro NVMe (1TB ×2)
Workloads separated to prevent contention:
| Drive | Role | Purpose |
|---|---|---|
| NVMe 1 | Cache Pool A | Docker containers + appdata |
| NVMe 2 | Cache Pool B | VM storage + AI models |
This layout should keep responsiveness consistent even when Plex, VMs, and AI services are active simultaneously.
Workload Architecture
Plex Media Server
Role: Media Processing Node
- Streaming
- Metadata processing
- Hardware transcoding (GPU planned)
Virtualization Layer
Role: VM Host Environment
Used for:
- Cybersecurity labs
- Testing environments
- Windows/Linux VMs
- Development workloads
The EPYC platform shines here due to thread availability.
File Server (NAS)
Role: Primary Storage Services
Provides:
- Network shares
- Media storage
- Backups
- Archive/ISO storage
NVMe cache accelerates frequently accessed data while bulk storage lives on the array.
Local AI Stack (OpenClaw + Llama)
Role: Local AI Inference Node
Components:
- OpenClaw gateway
- Ollama / Llama models
- Local vector memory
Running everything locally keeps data private and eliminates API costs while allowing experimentation with agents and automation workflows.
Logical Architecture
AI Services
↑
Plex | Virtual Machines | File Server
↑
AMD EPYC Compute Layer
↑
128 GB Memory Pool
↑
NVMe Cache Tier
↑
Storage Array
The goal is workload isolation while maximizing hardware utilization.
GPU Recommendations? (Main Question)
I’m currently trying to decide what GPU setup makes the most sense for this system and would really appreciate input from others running similar homelab or local AI environments.
Primary GPU Goals
- Accelerate local Llama inference (OpenClaw + Ollama)
- Handle Plex hardware transcoding
- Maintain good power efficiency for 24/7 operation
- Work reliably inside a server chassis
- Leave room for future multi-GPU expansion
Questions I’m Trying to Answer
- Is a single GPU enough for combined Plex + AI workloads?
- Does separating AI and Plex onto different GPUs actually help in practice?
- Is VRAM capacity more important than raw GPU speed for local LLMs?
- Enterprise cards vs consumer GPUs for long-term homelab reliability?
- What GPUs have worked best specifically with Ollama or local LLM setups?
Curious what others in the homelab or local AI space are running and what has held up well long term.