Hi
I was transferring some files from my Windows 10 PC over to to my unraid server when I got an error message from Windows telling me due to an i/o error the transfer had failed. I got an email notification from my server saying Drive 4 had been disabled and was now being emulated at the exact same time. I checked the unraid main page and I had a red error message saying
"disk 4 in error state (disc dsbl)"
And I also had a yellow error message saying
"Array has 2 discs with read errors"
I noticed on the unraid server main page Drive 4 had 1024 error's and also Drive 6 had over 9000 error's.
Drive 6 was still showing as online and did not have a red x next to it but all the data on that drive I've was missing when I checked through Windows Explorer.
I attempted a read-check before restarting the array and within a few minutes Drive 6 had 18 million error's so I stopped the check.
I downloaded the diagnostics and system log data and rebooted the array.
I had a look at the the diagnostic data and noticed unraid wasn't able to download any smart data from Drives 4 or 6.
After rebooting, Drive 6 was behaving normally with no errors and all data on the drive was accessible.
I have run an extended smart test on all drives which all completed.
This also happened once a few months ago in the exact same way, except I only had the problem with Drive 4, all the other drives were still operating normally without errors. Last time I just checked the data cables were connected properly then ran an extended smart self test on Drive 4. The smart test results seemed ok so I cleared the drive and rebuilt it just in case.
But now it's happened again I feel like I definitely need a second opinion on what I should do.
I should note that Drive 4 was the drive being written to in both cases, so I was thinking could it be possible that the drive and sata connection are fine and unraid took the drive that the data was supposed to go to offline due to an error elsewhere in my system?
I'm currently using turbo write mode so since all the drives need to be working in order for successful writes to happen I was thinking possibly if Drive 6 was the issue then that could have caused the write failure?
I am attaching the diagnostics and system log from before rebooting, and the smart extended self tests which I performed after reboot.
My drives are all seagate ironwolfs which are still under warranty.
If anybody could have have a little look and see whether they notice anything wrong with any of my smart test results or anything else that might look like the cause I would be very grateful
I'm already thinking about replacing the data cables with newer ones I have and changing which ports some of the drives are plugged into. But any other suggestions would be very welcome.
Many thanks
NAS-diagnostics-20200817-2156.zip
NAS-syslog-20200817-2056.zip
SMART extended self-tests After reboot.zip