I have a 6x2TB disk raidz2 ZFS pool which I am hosting in Solaris 11
My original M1015 (flashed to IT mode) installed 2 years ago when I built this server failed and was no longer recognized on the PCI-E bus. I replaced it with another M1015 yesterday (flashed to IT mode) and Solaris found all the disks again.
The ZFS pool however was put into SUSPENDED mode (probably due to the previous M1015 dying while running and all the disks vanishing before) and I see resilvering occurring on 2 disks (??), with all disks listed as unavailable.
I have no doubt there are some errors in the pool. But I have cleared (fmadm repaired, and zpool clear) the faults in hopes that the pool could be remounted in degraded state. However upon reboots the pool first comes up as DEGRADED (some disks show as unavailable, some as degraded) , then immediate transitions to SUSPENDED with all disks showing as unavailable and resilvering starts.
The resilvering speed starts at about 100MBps and rapidly ramps down to 50kbps or less. This equates to several hundred hours of expected resilvering time. What's more is that iostat shows ZERO transactions occurring on any of the disks in the pool.
I have no way to offline any of the disks or export the pool while it is in suspended state (also, I have no certain idea why it is entering suspended state when "fmadm faulty"entries are all reported as repaired)
Where have I gone wrong in replacing the SAS controller and how do I recover?
My original M1015 (flashed to IT mode) installed 2 years ago when I built this server failed and was no longer recognized on the PCI-E bus. I replaced it with another M1015 yesterday (flashed to IT mode) and Solaris found all the disks again.
The ZFS pool however was put into SUSPENDED mode (probably due to the previous M1015 dying while running and all the disks vanishing before) and I see resilvering occurring on 2 disks (??), with all disks listed as unavailable.
I have no doubt there are some errors in the pool. But I have cleared (fmadm repaired, and zpool clear) the faults in hopes that the pool could be remounted in degraded state. However upon reboots the pool first comes up as DEGRADED (some disks show as unavailable, some as degraded) , then immediate transitions to SUSPENDED with all disks showing as unavailable and resilvering starts.
The resilvering speed starts at about 100MBps and rapidly ramps down to 50kbps or less. This equates to several hundred hours of expected resilvering time. What's more is that iostat shows ZERO transactions occurring on any of the disks in the pool.
I have no way to offline any of the disks or export the pool while it is in suspended state (also, I have no certain idea why it is entering suspended state when "fmadm faulty"entries are all reported as repaired)
Where have I gone wrong in replacing the SAS controller and how do I recover?