Hello everybody,
I have been having issues with a nappit install on an esxi host and I do not know where to go from here.
Hardware config 1:
4x12tb sata drives + 4x1tb ssds on a Dell H310 in IT mode. Each 4 drives on one sff8087 to 4x sas connectors breakout cable.
HP Z420 mb, e52680v2, 128gb ddr3 1866 ecc.
HP z420 original 600watt psu with added sata power connectors.
ESXI 7, nappit vm.
4x12tb hdd -> raidz pool
4x1tb ssd -> raidz pool.
Hardware config 2:
4x14tb sas toshiba brand new drives + 4x1tb ssds on LSI 9207-8i 6Gbs SAS 2308 IT mode. Again, each 4 drives on its own breakout cable.
Gigabyte z590 vision D mb, i5 10400T/i7 11700F, 64 gb ddr4 non-ecc ram.
Antec earthwatts 450w 80+ platinum psu with added sata power connectors.
Esxi 7.0u3, nappit vm updated to latest available version (current).
4x12tb hdd -> raidz pool
4x1tb ssd -> raidz pool.
On my first hardware config one of my drives in one of my pools started to get errors while writing to that pool. This was always the last drive in the list. I suspected one of the 12tb drives was beginning to die. (these errors appeared on zpool status, but i really do not remember if they were on the write attribute or on cksum)
First I replaced the 12tb drive with a new one, and started rebuilding the pool just to find out that the new drive was starting to get same errors. So I switched back to the original hdd, and changed the ports on the HBA card. reseated all the cables in the drives, but still got errors on the same drive. So it was not bout to the HBA's port, or the connectors (I also switched drives so they would be connected to different ports)
Because of this I started to change the hardware, and I ended up with the second hardware config. I copied all the data to the new pool of new 14tb drives and all was ok for a few weeks until yesterday when I started getting errors on the all new configuration, on the last drive of the pool (when using zpool status). This baffles me. I do not know how or why this would happen.
on zpool status now I have:
raidz1-0 DEGRADED 0 0 0
c14t5000039B38809ECAd0 ONLINE 0 0 0
c15t5000039B3880A85Ad0 ONLINE 0 0 0
c16t5000039B3880FF4Ad0 ONLINE 0 0 0
c17t5000039B3880FDFEd0 FAULTED 0 0 0 too many errors
and on the web interface, under disks, I have:
What am I doing wrong?
I have been having issues with a nappit install on an esxi host and I do not know where to go from here.
Hardware config 1:
4x12tb sata drives + 4x1tb ssds on a Dell H310 in IT mode. Each 4 drives on one sff8087 to 4x sas connectors breakout cable.
HP Z420 mb, e52680v2, 128gb ddr3 1866 ecc.
HP z420 original 600watt psu with added sata power connectors.
ESXI 7, nappit vm.
4x12tb hdd -> raidz pool
4x1tb ssd -> raidz pool.
Hardware config 2:
4x14tb sas toshiba brand new drives + 4x1tb ssds on LSI 9207-8i 6Gbs SAS 2308 IT mode. Again, each 4 drives on its own breakout cable.
Gigabyte z590 vision D mb, i5 10400T/i7 11700F, 64 gb ddr4 non-ecc ram.
Antec earthwatts 450w 80+ platinum psu with added sata power connectors.
Esxi 7.0u3, nappit vm updated to latest available version (current).
4x12tb hdd -> raidz pool
4x1tb ssd -> raidz pool.
On my first hardware config one of my drives in one of my pools started to get errors while writing to that pool. This was always the last drive in the list. I suspected one of the 12tb drives was beginning to die. (these errors appeared on zpool status, but i really do not remember if they were on the write attribute or on cksum)
First I replaced the 12tb drive with a new one, and started rebuilding the pool just to find out that the new drive was starting to get same errors. So I switched back to the original hdd, and changed the ports on the HBA card. reseated all the cables in the drives, but still got errors on the same drive. So it was not bout to the HBA's port, or the connectors (I also switched drives so they would be connected to different ports)
Because of this I started to change the hardware, and I ended up with the second hardware config. I copied all the data to the new pool of new 14tb drives and all was ok for a few weeks until yesterday when I started getting errors on the all new configuration, on the last drive of the pool (when using zpool status). This baffles me. I do not know how or why this would happen.
on zpool status now I have:
raidz1-0 DEGRADED 0 0 0
c14t5000039B38809ECAd0 ONLINE 0 0 0
c15t5000039B3880A85Ad0 ONLINE 0 0 0
c16t5000039B3880FF4Ad0 ONLINE 0 0 0
c17t5000039B3880FDFEd0 FAULTED 0 0 0 too many errors
and on the web interface, under disks, I have:
c17t5000039B3880FDFEd0 | single | ok | 14 TB | S:0 H:252 T:3507 | TOSHIBA | MG07SCP14TE |
What am I doing wrong?