Hello everyone,
Really happy that the site and forum is back online!
I run into this issue a week ago and STH was not working anymore...so I felt a little lost
Anyways, I just added 6 Constellation ES3 SAS 4TB to my ZFS Server.
The drives are not NEW but they seems in a perfect condition and the warranty expires in 2018.
For some reasons just 2 drives are working...RED LED on the other 4.
I checked the drives in windows and seems everything ok, but no way in Solaris.
When I boot the system the onboard LSI controller recognize all the connected drives but once OmniOS is booted I see a "One or more I/O devices have been retired" message.
So I checked with fmadm faulty if something was wrong and that's the result:
It's weird since I get this message also if the drives are not warm or totally cold, so seems not related to temp...
Another thing I noticed is that the two working drives comes with an "A001" firmware and the others with a "0003" firmware. It's just a coincidence?
I sent a mail to the Seagate support, they don't know the differences between the two firmwares...! I really don't know what to do, Solaris doesn't like my enterprise drives...What do you guys think? I should RMA the not working drives? Thanks in advance.
PS: I'm also having hard time to read the smart info of the retired drives.
Really happy that the site and forum is back online!
I run into this issue a week ago and STH was not working anymore...so I felt a little lost
Anyways, I just added 6 Constellation ES3 SAS 4TB to my ZFS Server.
The drives are not NEW but they seems in a perfect condition and the warranty expires in 2018.
For some reasons just 2 drives are working...RED LED on the other 4.
I checked the drives in windows and seems everything ok, but no way in Solaris.
When I boot the system the onboard LSI controller recognize all the connected drives but once OmniOS is booted I see a "One or more I/O devices have been retired" message.
So I checked with fmadm faulty if something was wrong and that's the result:
Code:
--------------- ------------------------------------ -------------- ---------
TIME EVENT-ID MSG-ID SEVERITY
--------------- ------------------------------------ -------------- ---------
Jun 06 23:40:00 e50e15b5-19d4-49d2-f1df-a7c715df3106 DISK-8000-12 Major
Host : server
Platform : X9SRH-7F-7TF Chassis_id : 0123456789
Product_sn :
Fault class : fault.io.disk.over-temperature
Affects : dev:///:devid=id1,sd@n5000c50057b034b3//pci@0,0/pci8086,e08@3/pci15d9,691@0/iport@f0/disk@w5000c50057b034b2,0
faulted and taken out of service
FRU : "Slot 09" (hc://:product-id=LSI-SAS2X36:server-id=:chassis-id=5003048000b4aa3f:serial=Z1Z2G3X10000C416CR4X:part=SEAGATE-ST4000NM0023:revision=0003/ses-enclosure=0/bay=8/disk=0)
faulty
Description : A disk's temperature exceeded the limits established by its manufacturer.
Refer to DISK-8000-12 for more information.
Response : None.
Impact : Performance degradation is likely and continued disk operation beyond the temperature threshold can result in disk damage and potential data loss.
Action : Ensure that the system is properly cooled, that all fans are functional, and that there are no obstructions of airflow to the affected disk.
It's weird since I get this message also if the drives are not warm or totally cold, so seems not related to temp...
Another thing I noticed is that the two working drives comes with an "A001" firmware and the others with a "0003" firmware. It's just a coincidence?
I sent a mail to the Seagate support, they don't know the differences between the two firmwares...! I really don't know what to do, Solaris doesn't like my enterprise drives...What do you guys think? I should RMA the not working drives? Thanks in advance.
PS: I'm also having hard time to read the smart info of the retired drives.