Hi STH community,
I just wanted to give everyone a quick heads-up that we may be experiencing a small bit of downtime.
What is happening?
We are moving from our current 1/4 cabinet of colocation excellence to a 1/2 cabinet with more power, more bandwidth and more space.
During the move, we are combining some of the existing hardware with new hardware and that will necessitate physically moving boxes and a short period of downtime. The other options cost significantly more and this is simply one cabinet to another so it should not take "too" long. (fingers crossed.)
When
I got the e-mail that they are working on this over the next few days. So it will likely happen soon.
More details
After the "great crash" of 2014 the architecture needed to change (significantly.) We are moving to much higher end SSDs from Intel (S3500 and S3700), SanDisk (Lightning SLC drives and Optimus), Seagate (Pulsar.2) all in RAID 1 arrays with hot spares everywhere.
We are adding another backup node in the form of a 4U server that will have multiple RAID 1 volumes and hot spares. Currently this is setup as a ZFS-based backup solution, but we do have SLC and MLC SAS drives available for use.
We are adding more nodes/ chassis to provide a small lab environment, additional test nodes and etc. The great crash seems to have been due to a power issue in a 4-in-1 chassis while we were using 3 of the 6 total nodes we have in the colocation during setup of the second chassis. We are moving to more chassis so we never have to do maintenance where we have all of our eggs in one basket again. Lesson learned.
10/40GbE is being added to help cope with the extra traffic (1GbE will be there just in case too.)
Conclusion
The bottom line is that this just needs to get done. Maintenance on the current setup leaves us with only one functional chassis and I want to get our setup to something that has room to breathe. Costs are relatively low so it made sense.
I just wanted to give everyone a quick heads-up that we may be experiencing a small bit of downtime.
What is happening?
We are moving from our current 1/4 cabinet of colocation excellence to a 1/2 cabinet with more power, more bandwidth and more space.
During the move, we are combining some of the existing hardware with new hardware and that will necessitate physically moving boxes and a short period of downtime. The other options cost significantly more and this is simply one cabinet to another so it should not take "too" long. (fingers crossed.)
When
I got the e-mail that they are working on this over the next few days. So it will likely happen soon.
More details
After the "great crash" of 2014 the architecture needed to change (significantly.) We are moving to much higher end SSDs from Intel (S3500 and S3700), SanDisk (Lightning SLC drives and Optimus), Seagate (Pulsar.2) all in RAID 1 arrays with hot spares everywhere.
We are adding another backup node in the form of a 4U server that will have multiple RAID 1 volumes and hot spares. Currently this is setup as a ZFS-based backup solution, but we do have SLC and MLC SAS drives available for use.
We are adding more nodes/ chassis to provide a small lab environment, additional test nodes and etc. The great crash seems to have been due to a power issue in a 4-in-1 chassis while we were using 3 of the 6 total nodes we have in the colocation during setup of the second chassis. We are moving to more chassis so we never have to do maintenance where we have all of our eggs in one basket again. Lesson learned.
10/40GbE is being added to help cope with the extra traffic (1GbE will be there just in case too.)
Conclusion
The bottom line is that this just needs to get done. Maintenance on the current setup leaves us with only one functional chassis and I want to get our setup to something that has room to breathe. Costs are relatively low so it made sense.