Hard drive disabled during file transfer again, Seeking Second Opinion


Recommended Posts

Hi

I was transferring some files from my Windows 10 PC over to to my unraid server when I got an error message from Windows telling me due to an i/o error the transfer had failed. I got an email notification from my server saying Drive 4 had been disabled and was now being emulated at the exact same time. I checked the unraid main page and I had a red error message saying
"disk 4 in error state (disc dsbl)"
And I also had a yellow error message saying
"Array has 2 discs with read errors"

I noticed on the unraid server main page Drive 4 had 1024 error's and also Drive 6 had over 9000 error's.

Drive 6 was still showing as online and did not have a red x next to it but all the data on that drive I've was missing when I checked through Windows Explorer.

I attempted a read-check before restarting the array and within a few minutes Drive 6 had 18 million error's so I stopped the check.

I downloaded the diagnostics and system log data and rebooted the array.

I had a look at the the diagnostic data and noticed unraid wasn't able to download any smart data from Drives 4 or 6.

After rebooting, Drive 6 was behaving normally with no errors and all data on the drive was accessible.

I have run an extended smart test on all drives which all completed.

This also happened once a few months ago in the exact same way, except I only had the problem with Drive 4, all the other drives were still operating normally without errors. Last time I just checked the data cables were connected properly then ran an extended smart self test on Drive 4. The smart test results seemed ok so I cleared the drive and rebuilt it just in case.

But now it's happened again I feel like I definitely need a second opinion on what I should do.

I should note that Drive 4 was the drive being written to in both cases, so I was thinking could it be possible that the drive and sata connection are fine and unraid took the drive that the data was supposed to go to offline due to an error elsewhere in my system?

I'm currently using turbo write mode so since all the drives need to be working in order for successful writes to happen I was thinking possibly if Drive 6 was the issue then that could have caused the write failure?



I am attaching the diagnostics and system log from before rebooting, and the smart extended self tests which I performed after reboot.

My drives are all seagate ironwolfs which are still under warranty.

If anybody could have have a little look and see whether they notice anything wrong with any of my smart test results or anything else that might look like the cause I would be very grateful :)

I'm already thinking about replacing the data cables with newer ones I have and changing which ports some of the drives are plugged into. But any other suggestions would be very welcome.

Many thanks :) 


 

NAS-diagnostics-20200817-2156.zip NAS-syslog-20200817-2056.zip SMART extended self-tests After reboot.zip

Link to comment
2 hours ago, aukaauka said:

could it be possible that the drive and sata connection are fine and unraid took the drive that the data was supposed to go to offline due to an error elsewhere in my system?

no Unraid disables a disk when a write to it fails. Simple as that.

 

Disks SMART looks OK. Are these on the same controller?

  • Thanks 1
Link to comment

Both disks share the same Asmedia controller and dropped at the same time:

 

Aug 17 21:17:17 NAS kernel: ata8: softreset failed (1st FIS failed)
Aug 17 21:17:17 NAS kernel: ata8: reset failed, giving up
Aug 17 21:17:17 NAS kernel: ata8.00: disabled
Aug 17 21:17:17 NAS kernel: ata8: EH complete
Aug 17 21:17:17 NAS kernel: ata7: softreset failed (1st FIS failed)
Aug 17 21:17:17 NAS kernel: sd 8:0:0:0: [sdh] tag#1 UNKNOWN(0x2003) Result: hostbyte=0x04 driverbyte=0x00
Aug 17 21:17:17 NAS kernel: sd 8:0:0:0: [sdh] tag#1 CDB: opcode=0x35 35 00 00 00 00 00 00 00 00 00
Aug 17 21:17:17 NAS kernel: print_req_error: I/O error, dev sdh, sector 0
Aug 17 21:17:17 NAS kernel: ata7: reset failed, giving up
Aug 17 21:17:17 NAS kernel: ata7.00: disabled

 

Asmedia controllers usually work fine with Unraid, but there could be an actual problem with it, that or the disks for example share a power splitter/cable with issues.

  • Thanks 1
Link to comment

Just checked my motherboard manual and it has

 

6 x SATA3 6.0 Gb/s Connectors by Intel® Z170

4 x SATA3 6.0 Gb/s Connectors by ASMedia ASM1061

So does that mean I have 2 sata controllers, rather than say each set of parallel ports being separate?

Two of the Intel ports are disabled because I have an nvme Drive installed.

Drives 3, 4 and 6 are plugged into Asmedia ports but drives 4 and 6 are in ports stacked on top of each other.

As for power, drives 4 and 6 as well as drive 1 and parity all share a sata Type 4 PSU power cable.

I've had the system for a couple of years now but at the beginning of the year I did add a couple of extra drives and rearranged some of the sata cables so possibly I didn't have much plugged into the Asrock ports beforehand so maybe that's why I'm only starting to have problems only now?

Link to comment
16 minutes ago, johnnie.black said:

There are 3 controllers, 6 ports Intel controller plus two 2 port Asmedia controllers.

 

Since you still have Intel ports available use those first.

Interesting, ok so that would suggest I'm possibly having an issue with one of the Asmedia controllers? (The one which which drives 4 and 6 are plugged into)

Apologies I should have clarified that all my Intel SATA ports either in use or disabled because of the nvme drive I have installed

Link to comment

Alright so I could try plugging drive 4 into the other Asmedia controller which has 1 SATA Port free and see if i have any more problems.

 

If that doesn't help then then I guess i'll need to get a PCIe SATA card

 

Are there any types of SATA card that you would recommend I should or shouldn't get?

Edited by aukaauka
Link to comment

Brilliant, thank you so much for all the help. 😄

5 ports should be great and they don't seem that expensive. For now I've moved some of the sata cables around but if I have any more problems I'll buy a sata card.


Also I just thought of an unrelated nas question, all my hard drives are in an aluminium PC case with no vibration dampening and the Reverberation through the case is quite loud.
Seagate state that ironwolf drives are good to be in enclosures of up to 8 drives, but I've been wondering recently if it would be better for the drives if i got a case where the hard drive mounting system means they are better vibration isolated from each other?

Many thanks :)

 

Link to comment

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.