Anyone out there have any experience with really big (say 24 disk) SSD arrays? With SSD prices on a downward trend, I would expect to see more ideas like the Dirt Cheap Data Warehouse (DCDW) showing up... but I'm either looking in the wrong places, or no one is really doing this. Or they're doing it but not talking about it!
We currently have a big iron filer backing a research/simulation platform. It's 99.9% reads, effectively a WORM workload. The reads are sequential, but there are hundreds running in parallel, across roughly 30 TB of data.
The DCDW got me thinking that we can probably build our own consumer SSD-based system for considerably less than the big name vendors are charging. I'm thinking something along the lines of a 24x 2.5" chassis (e.g. Supermicro SC216) and filling that with 2TB Samsung EVO drives. I haven't yet decided on which CPU(s), motherboard, how much RAM or even the right HBA/RAID controller to use (suggestions welcome!). I'm actually thinking I might be able to get away with software RAID. Two 12-disk RAID-6 arrays stripped with RAID-0 (RAID-60) seems like a good starting point for experimentation.
If such a system could support a decent number of clients (say 10+ compute nodes), then I could just build a bunch of these DCDW-type systems and replicate the data across them. Kind of a poor man's cluster.
Just throwing this out there to see if anyone has any experience with anything similar they're willing to share. Any thoughts on gotchas or potential pitfalls or ideas in general are also welcome!
Thanks!
We currently have a big iron filer backing a research/simulation platform. It's 99.9% reads, effectively a WORM workload. The reads are sequential, but there are hundreds running in parallel, across roughly 30 TB of data.
The DCDW got me thinking that we can probably build our own consumer SSD-based system for considerably less than the big name vendors are charging. I'm thinking something along the lines of a 24x 2.5" chassis (e.g. Supermicro SC216) and filling that with 2TB Samsung EVO drives. I haven't yet decided on which CPU(s), motherboard, how much RAM or even the right HBA/RAID controller to use (suggestions welcome!). I'm actually thinking I might be able to get away with software RAID. Two 12-disk RAID-6 arrays stripped with RAID-0 (RAID-60) seems like a good starting point for experimentation.
If such a system could support a decent number of clients (say 10+ compute nodes), then I could just build a bunch of these DCDW-type systems and replicate the data across them. Kind of a poor man's cluster.
Just throwing this out there to see if anyone has any experience with anything similar they're willing to share. Any thoughts on gotchas or potential pitfalls or ideas in general are also welcome!
Thanks!