DanW

Members
  • Posts

    34
  • Joined

  • Last visited

Everything posted by DanW

  1. I've checked all cables and they're connected firmly. Disk is rebuilding currently, I think I'm gonna replace the SAS controller as a precaution. Too many weird disk issues in the last few months.
  2. Hey, thanks @JorgeB I was thinking this but I replaced all the SAS cables with higher quality lock in cables and double checked power connections a couple months ago, uptime is 1 month 15 days without issues. I'll give it a reboot and see what happens.
  3. Seems I also cannot get my array to stop.
  4. Hey all, Would someone be able to advise me please. Is this a sign of a failing disk or some weird SAS controller bug? I had two drives fail a couple of months ago and now i'm thinking either my SAS controller is the issue or i'm really unlucky 😂 Thank you for any help or suggestions. dansunraidnas-diagnostics-20230315-1817.zip
  5. I've noticed something weird. Seems to be stuck like this.
  6. Really good suggestions, thank you! My appdata is set to use cache only, not sure why I have some "bytes" in the array? Domains is just a backup of the VM vdisks I have running on the unassigned NVME drives so I dont have it in cache "only". Unfortunately I am making use of one the unassigned SSDs right now and have plans for the other unassigned SSD. I don't know why I had not set system to cache only, I have done this now but probably need to move the files back to cache.
  7. Oops, here's the correct diagnostics 🙂 dansunraidnas-diagnostics-20230127-2050.zip
  8. So everything is back to normal 🎉 Thank you everyone for the help and support. dansunraidnas-diagnostics-20230127-1851.zip
  9. Apologies, please see attached. Rebuild of disk 8, the second disk to fail, is still underway. The array is live, the data that originally disappeared (when the first error with disk 7 occurred) is back, which is really positive. Disk 8 was emulated immediately when it failed so I didn't notice any data loss the second time. dansunraidnas-diagnostics-20230125-2349.zip
  10. Disk 8 is rebuilding, no other issues so far.
  11. Thank you for your help 🙂 I really appreciate your knowledge and suggestions. I'm going to go ahead and replace disk 8 now and rebuild it hopefully without any more issues 🤞
  12. Interesting, is that an issue?
  13. I haven't replaced disk 8 yet (I replaced disk 7 first as it was the initial problem disk with filesystem corruption), I'm going to change some SAS cables and insert a new drive to replace disk 8 then start the rebuild soon. I will provide updates.
  14. Really? I've replaced the disk (disk 7) and the disk has been rebuilt without any issues. Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... - scan filesystem freespace and inode maps... - found root inode chunk Phase 3 - for each AG... - scan (but don't clear) agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - agno = 4 - agno = 5 - agno = 6 - agno = 7 - agno = 8 - agno = 9 - agno = 10 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - agno = 4 - agno = 5 - agno = 6 - agno = 7 - agno = 8 - agno = 9 - agno = 10 No modify flag set, skipping phase 5 Phase 6 - check inode connectivity... - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify link counts... No modify flag set, skipping filesystem flush and exiting. I've got to replace disk 8 now as it failed during the rebuild of disk 7, lucky I had two parity drives.
  15. I didn't attempt the recovery, I put a new drive in to replace this drive. Shortly after starting recovery, disk 8 reported an I/O error too and has been disabled. These drives are old and have a lot of uptime but seems to be a strange coincidence that they would both die together so soon. To rule out heat issues I've pointed fans at my SAS devices. I've also ordered some higher quality SAS cables. Going to be keeping an eye on the SAS controller & HBA, it has been fine for months and my drives are old, so could just be a coincidence. I'm currently using the following SAS devices: IBM SAS HBA M1015 IT Mode 6Gbps PCI-e 2.0 x8 LSI 9220-8i Intel 24 port 6 Gb/s SATA SAS RAID Expander Card PBA E91267-203 RES2SV240
  16. I got this when running the xfs_repair -v command. I didn't see anything about this in the instructions so i have no idea what to do next. Im going to just remove the drive and drop the new one in.
  17. I think I'm going to attempt to recover the file system then replace the drive later today.
  18. Have you changed the SATA data cable? Could be a bad sata cable or something wrong with the SATA port on the motherboard. Especially if a new drive dropped in its place does the same thing, it's unlikely you have two bad drives in a row.
  19. Hey everyone, I started getting I/O errors on one of my drives. Noticed a load of my files suddenly disappear in my shares and went straight to the system log to see what was going on. I've ran the check on all 13 drives in maintenance mode and it's just the one playing up (disk 7) from what I can see. Any recommendations? Just run the check without -nv and see if it recovers the drive? I have two parity drives and I have a new drive spare that I could drop in to replace it. Some advice from someone who has experience in this area would be greatly appreciated, thank you check-nv.txt dansunraidnas-diagnostics-20230123-2257.zip
  20. Hey @MrX, this is really good news, thank you for checking in and confirming. I look forward to the stable release of 6.10
  21. I haven't tried the 6.10 RC although I have been trying to find a changelog or something that might point to issues with AMD CPUs being resolved. All I've been able to find so far shows that there have been some onboard GPU improvements for AMD and Intel. So I'm not sure if 6.10 is going to fix the issue either.
  22. 3 months later and I still haven't figured this out. I've tried passing through a USB with the pfSense installer on it. I've upgraded to the latest MOBO BIOS. Maybe an unraid update will fix this?
  23. I'm on 6.9.2 and noticed this on two of my UNRAID servers after noticing a lot of heat and fan noise. One is a Ryzen 5 5600X and the other is a Xeon 1270 v2, the -i br0 or -i bond0 depending on your network configuration does seem to resolve the issue. This really needs to be fixed, there are going to be a lot of UNRAID servers around the world unnecessarily draining power and generating heat because of this issue.