I've been running a home lab environment for over 15 years. I've had several devices in use during those years. However, it's been a while since I was in "hardware" mode, so I'm _very_ behind the times. I was hoping that my ADHD super power (believing I can master a skill in 24 hours, but really can't) would help me, and there's a _ton_ I've missed out on in the past 8 years. Here's my basic break down of current devices:
NAS01:
ESXi01 (primary):
Dell R210 II (pfSense CE 2.7.2 bare metal):
Here are the docker workloads that I'm currently running, and a couple that I'd like to add.
Here are the VMs I'm currently running:
It looks like a lot, but I believe in reality, I don't have much workload to consider. All of these servers are connected via 1Gb ethernet to a Cisco 3750 PoE+ Switch Stack with 1KW PSUs. That switch stack is running as a L3 switch with a transit network to pfSense. They are using OSPF to route. The switch is also an mDNS reflector for Roon, Roon Endpoints, and other IoT devices on the various VLANs.
What are the issues?
I've moved my main home entertainment system to 4K. In the process of obtaining 4K Linux ISOs, I've had to limit access to my 4K library to only internal users due to horrible ISP upload. The current desktop runs Plex and isn't beefy enough to transcode my 4K Linux ISOs. This has caused some lower WAF when she tries to view a show/movie on a non-4K TV. I don't want to keep several copies of Linux Distro just to accomodate "ease of viewing".
The other issue I have is power usage. The NAS, and all the other systems are pulling a lot of power. I don't have an estimate at this time as I've never really kept track of power usage. I'm sure the NAS is the most power hungry due to all the drives in there, and those drives won't change. I can tell you we have a high power bill.
What is the solution?
If I can consolidate all these "hosting" servers into a smaller cluster (for HA purposes) I am hoping I can save a lot on power/heat, and noise. I am also hoping I can upgrade the CPU/MOBO/RAM in the NAS as well to accomodate the massive increase in storage I've evolved into. Here's a diagram of a proposed solution:

The first big change here has to do with adding a new Layer 3 switch. This switch would be a 40Gb switch that can also do DHCP and static DHCP. However, it also needs to do some directed broadcast magic as well. Roon requires this. I'm still in the early stages of home work on the switch, but I think it should be possible.
The next big change is the use of 3 new Proxmox hosts in a cluster with Ceph as the shared storage. This is also brand new to me. I'm hoping that this configuration will allow me to have high availability of LXCs and VMs (specifically pfSense and Home Assistant). I work from home full time, and the wife WFH's a few days a week, so up time is critical.
I need help, please
Here's where I need a ton of help. I'm trying to decide if I've got the right idea behind a Proxmox cluster and what hardware I should get to accomplish my goals. Here are my thoughts right now:
NAS01:
This needs a new mobo/CPU/RAM. I'm thinking about:
Supermicro MBD MBDH12SSLiO
AMD Epyc 7F52
Noctua CPU Cooler NH-D9 TR5-SP6
128GB ECC 2666 RAM
Mellanox MCX455A-ECAT ConnectX-4 EDR
PVE Hosts:
This is the area that I'm really not sure what to do. I think my hang up is on disk storage. In the diagram above, I've got HDDs in a Ceph pool. The thought here was that if I have some large files I need to store (camera recordings, Pingvin Share files, temp downloads, etc) I could do that easily on the HDD Ceph pool. This frees up "expensive" drive slots in the NAS. To accomodate these larger HDDs, I'll need a larger case. The larger case means I could also install the same Mellanox card as above, and an Intel Arc A380 in each host. Hopefully, the dual Coral TPU as well. I was honestly just thinking about using the same Supermicro build above. But that comes to like $2500, and I'd need 3 for the cluster. I could really use some opinions on this host config or others that can accomodate 2 boot drives (SATA DOM), 2 NVME 4TB drives, and 2 or 4 16TB HDDs. I also need a case! The 45 Home Lab HL15 case is $1100!! I was thinking about the MS-O1, but I think the Intel Big Little CPU architecture, lack of space for HDDs, and no room for the Mellanox NIC make them a non-starter.
Network Switch:
Again, another area where I need to do a lot of homework. I think an L3 40Gb switch is totally doable. It's the directed broadcast and possibly any license changes needed to accomodate those things.
I'm sure I'm overthinking this whole ordeal. I just don't know what to do about my Ceph pools for the VMs. I'd really appreciate anyone's input and experience on these topics. Thanks for reading!
NAS01:
- Chenbro NR40700 48 bay enclosure with backplane
- SuperMicro Motherboard (don't remember model)
- Intel(R) Xeon(R) CPU E5-2603 0 @ 1.80GHz
- 32GB RAM
- 150+ TB of disks in various ZFS RAIDZ2 or MIRRORs
ESXi01 (primary):
- Dell R430
- Intel(R) Xeon(R) CPU E5-2690 v3 @ 2.60GHz
- 96GB RAM
- 1.25TB storage
- 2 x 1Gb NICs
- 1 x LSI 2308 Passthrough to VM for Tape Library access in Bareos Storage Daemon
- Lenovo Tiny
- Intel(R) Core(TM) i5-4570T CPU @ 2.90GHz
- 16GB RAM
- 256GB storage
- 1Gb NIC
- Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz
- 32GB RAM
- 256GB ROOT
- 512GB M2 NVME for /home/user
- 2 x 256GB ZFS MIRROR for docker volume storage
- 1 x 1Gb NIC
- Intel(R) Core(TM) i7-8700 CPU @ 3.20GHz
- Supermicro X11SCZ-F
- 8GB RAM
- 256GB ROOT
- 10TB camera recording storage
- 1 x 1Gb NIC
Dell R210 II (pfSense CE 2.7.2 bare metal):
- Intel(R) Xeon(R) CPU E3-1240 V2 @ 3.40GHz
- 8 GB RAM
- 2 x 256GB SSD in ZFS MIRROR
- 2 x 1Gb NIC (one WAN, one LAN)
Here are the docker workloads that I'm currently running, and a couple that I'd like to add.
Plex | Running on old 4000 series Intel. Will need 4K transcode options, HOST GPU?? | |
Jellyfin?? | FUTURE. Will need 4K transcode options, HOST GPU?? | |
Frigate | FUTURE. Can it share the GPU with HOST and PLEX/JF? Coral TPU? | |
Immich | FUTURE. Will need 4K transcode options, HOST GPU?? | |
Jackett | ||
2 x RADARR | ||
2 x SONARR | ||
LIDARR | ||
READARR | ||
Roon | ||
Roon Extension Manager | ||
LubeLogger | ||
Bareos DB (Maria SQL) | ||
Bareos Director | ||
Bareos WebUI | ||
Bareos Storage Daemon | Move this to TrueNAS Scale and use Incus. Install the LSI2308 directly into NAS01 for Tape Library access. | |
Beets | ||
Calibre-Web | ||
Pingvin Share | ||
smtpd | ||
tftp-server | ||
transmission | ||
Unifi?? | ||
vlmcsd | ||
ZNC | ||
AVAHI for mDNS between VLANs | FUTURE | |
Minecraft Server | FUTURE | |
2 x SyncThing | Currently running on TrueNAS Core | |
Gitea | FUTURE | |
FreeRADIUS | FUTURE | |
Certificates | FUTURE |
Here are the VMs I'm currently running:
pfSense | Passthrough AES device? | |
Linux Workstation to replace Desktop w/ 3 monitors | Another dedicated GPU? | |
Linux Backup Server | CrashPlan runs here | |
Home Assistant | Will need POE Zwave, POE Matter | |
Windows 2016 | ||
Windows 2019 | Active Directory | |
Home Desktop Replacement | FUTURE, DEDICATED GPU? This isn't a necessity. |
It looks like a lot, but I believe in reality, I don't have much workload to consider. All of these servers are connected via 1Gb ethernet to a Cisco 3750 PoE+ Switch Stack with 1KW PSUs. That switch stack is running as a L3 switch with a transit network to pfSense. They are using OSPF to route. The switch is also an mDNS reflector for Roon, Roon Endpoints, and other IoT devices on the various VLANs.
What are the issues?
I've moved my main home entertainment system to 4K. In the process of obtaining 4K Linux ISOs, I've had to limit access to my 4K library to only internal users due to horrible ISP upload. The current desktop runs Plex and isn't beefy enough to transcode my 4K Linux ISOs. This has caused some lower WAF when she tries to view a show/movie on a non-4K TV. I don't want to keep several copies of Linux Distro just to accomodate "ease of viewing".
The other issue I have is power usage. The NAS, and all the other systems are pulling a lot of power. I don't have an estimate at this time as I've never really kept track of power usage. I'm sure the NAS is the most power hungry due to all the drives in there, and those drives won't change. I can tell you we have a high power bill.
What is the solution?
If I can consolidate all these "hosting" servers into a smaller cluster (for HA purposes) I am hoping I can save a lot on power/heat, and noise. I am also hoping I can upgrade the CPU/MOBO/RAM in the NAS as well to accomodate the massive increase in storage I've evolved into. Here's a diagram of a proposed solution:

The first big change here has to do with adding a new Layer 3 switch. This switch would be a 40Gb switch that can also do DHCP and static DHCP. However, it also needs to do some directed broadcast magic as well. Roon requires this. I'm still in the early stages of home work on the switch, but I think it should be possible.
The next big change is the use of 3 new Proxmox hosts in a cluster with Ceph as the shared storage. This is also brand new to me. I'm hoping that this configuration will allow me to have high availability of LXCs and VMs (specifically pfSense and Home Assistant). I work from home full time, and the wife WFH's a few days a week, so up time is critical.
I need help, please
Here's where I need a ton of help. I'm trying to decide if I've got the right idea behind a Proxmox cluster and what hardware I should get to accomplish my goals. Here are my thoughts right now:
NAS01:
This needs a new mobo/CPU/RAM. I'm thinking about:
Supermicro MBD MBDH12SSLiO
AMD Epyc 7F52
Noctua CPU Cooler NH-D9 TR5-SP6
128GB ECC 2666 RAM
Mellanox MCX455A-ECAT ConnectX-4 EDR
PVE Hosts:
This is the area that I'm really not sure what to do. I think my hang up is on disk storage. In the diagram above, I've got HDDs in a Ceph pool. The thought here was that if I have some large files I need to store (camera recordings, Pingvin Share files, temp downloads, etc) I could do that easily on the HDD Ceph pool. This frees up "expensive" drive slots in the NAS. To accomodate these larger HDDs, I'll need a larger case. The larger case means I could also install the same Mellanox card as above, and an Intel Arc A380 in each host. Hopefully, the dual Coral TPU as well. I was honestly just thinking about using the same Supermicro build above. But that comes to like $2500, and I'd need 3 for the cluster. I could really use some opinions on this host config or others that can accomodate 2 boot drives (SATA DOM), 2 NVME 4TB drives, and 2 or 4 16TB HDDs. I also need a case! The 45 Home Lab HL15 case is $1100!! I was thinking about the MS-O1, but I think the Intel Big Little CPU architecture, lack of space for HDDs, and no room for the Mellanox NIC make them a non-starter.
Network Switch:
Again, another area where I need to do a lot of homework. I think an L3 40Gb switch is totally doable. It's the directed broadcast and possibly any license changes needed to accomodate those things.
I'm sure I'm overthinking this whole ordeal. I just don't know what to do about my Ceph pools for the VMs. I'd really appreciate anyone's input and experience on these topics. Thanks for reading!