Jump to content
LAST CALL on the Unraid Summer Sale! 😎 ⌛ ×

Two disks "died" after 6.12 upgrade


Go to solution Solved by ericswpark,

Recommended Posts

Hi everyone, after upgrading 6.11.5 -> 6.12.0 I was greeted with this lovely sight:

 

image.png.491616c052ba429b88891f6e7c29b6c9.png

 

Thank god for dual parity (and a separate backup server). Not worrying too much about the data really takes the strain off of surprises like these.

 

However, I suspect that the drives aren't actually "dead", since they were working fine right up to the upgrade. Also, it's rather strange for two drives to "die" like that. I suspect the cable that plugs into those two drives may have become loose, either on the drive end or on the LSI card's end.

 

Unfortunately, the server is located remotely and therefore I cannot go and check it physically. I was wondering if anybody could find any clues as to why the drives are not coming up in the logs. I did find some messages saying "SATA link down", but it didn't say why, or I might've missed it. If not I'll have to check it over the next time I get a chance to inspect it in person. Any ideas are appreciated! Diagnostics attached.

 

 

dipper-diagnostics-20230616-0845.zip

Link to comment

I happened to check the webGUI today and noticed that the parity drive... came back on its own? The syslogs just say a device connect or power on event occurred. The drive somehow just woke back up and decided to work normally.

 

The other data disk is still missing, but I'm hoping that with enough power fed to it it would eventually recover itself like the parity drive.

 

Still very curious as to what's going on. It doesn't seem like a loose connection at all as the UDMA CRC values are all still at zero for the parity drive that returned. ¯\_(ツ)_/¯

Link to comment
  • 4 weeks later...
  • Solution

After upgrading to 6.12.3 today I found that the same two drives had died again. A physical inspection last time didn't turn up anything, but I decided to check again. I noticed that the drives that had "died" were connected to my HBA with one of those SAS to SATA cables, and the cable on the SATA end had gotten a bit bent as I built the NAS in a mini-ITX case.

 

I replaced the entire cable and it seems like the problem has been fixed? I'll keep the old cable around, but as long as two drives don't drop out during upgrades I think it's safe to rule this as a cable issue. The missing drives didn't even show up in the SAS configuration utility when the suspected faulty cable was used.

 

moral of the story: change cables and don't build your NAS in a mini-ITX case

  • Like 1
Link to comment

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...