Jump to content

JorgeB

Moderators
  • Posts

    67,793
  • Joined

  • Last visited

  • Days Won

    708

Everything posted by JorgeB

  1. So far I haven't found any corruption signs on Dell Servers with these NICs, so it *might* be safe to use vt-d, see below how you can do that: https://forums.unraid.net/bug-reports/stable-releases/after-i-upgraded-to-6102-my-network-interfaces-dissappeared-r1956/?do=findComment&comment=19584
  2. You need to either disable vt-d or unblacklist the NIC, Dell servers appear to not have the corruption issue, so it *might* be OK to use the NIC, you can to that by doing this: https://forums.unraid.net/bug-reports/stable-releases/after-i-upgraded-to-6102-my-network-interfaces-dissappeared-r1956/?do=findComment&comment=19584
  3. If you don't have a NIC that uses the tg3 driver you can, if you have one it might be safer for now to leave disable, at least until we get to the bottom of exactly what causes the problem.
  4. bnx2 is fine, even tg3 with Dell might be fine, at least didn't find any issues with one so far.
  5. You don't have a NIC the uses the tg3 driver, so unrelated to this, could be a network config issue.
  6. You can also just disable vt-d by adding 'intel_iommu=off' to syslinux.cfg, see how to here.
  7. As you mention the GPUs are being detected, you should post this in the Nvidia plugin support thread.
  8. If you have doubts how to check if your hardware is affected or workaround this please post the diagnostics.
  9. With v6.9.x it won't even ask you to create a password, and can't see how it would ask for the old password after just copying the key.
  10. That is a 1st gen Seagate SMR drive, never had one but remember reading write performance isn't great, 2nd gen Seagate drives like the ST8000AS0002 usually perform well enough with Unraid, the ones I have are about as fast as normal CMR drives for most cases, Toshiba and WD SMR drives (and possibly other Seagate models) don't perform as well, I have an array with Toshiba drives that frequently slows down to 5MB/s for writes.
  11. Do you need a HBA for the 15 drives or will you also be using onboard SATA?
  12. All should work the same, assuming they are flash to IT mode.
  13. Already replied in the other thread you posted, please don't post about the same thing in multiple threads, and like mentioned we need the diagnostics to see the hardware, they can be from v6.10.1
  14. It can be, since vt-d is now disable delete and re-create the docker image and monitor to make sure no more corruptions are found.
  15. Please post the diagnostics, they can be from v6.10.1
  16. I would suggest creating a new bug report in the stable releases section, and don't forget the diagnostics.
  17. Cache device dropped offline: May 28 03:35:21 uNAS kernel: ata1: hard resetting link May 28 03:35:27 uNAS kernel: ata1: link is slow to respond, please be patient (ready=0) May 28 03:35:56 uNAS kernel: ata1: COMRESET failed (errno=-16) May 28 03:35:56 uNAS kernel: ata1: limiting SATA link speed to 3.0 Gbps May 28 03:35:56 uNAS kernel: ata1: hard resetting link May 28 03:36:01 uNAS kernel: ata1: COMRESET failed (errno=-16) May 28 03:36:01 uNAS kernel: ata1: reset failed, giving up May 28 03:36:01 uNAS kernel: ata1.00: disabled May 28 03:36:01 uNAS kernel: ata1: EH complete Check/replace cables.
  18. And this posted above doesn't work for you? https://forums.unraid.net/topic/119502-bzimage-checksum-error/?do=findComment&comment=1108354
  19. Taking a more careful look the problem is that you started a balance, and since it wasn't able to finish because of the no space issue it's resuming after rebooting: May 28 09:49:37 Tower kernel: BTRFS info (device sdo1): balance: resume -dusage=90 -musage=90 -susage=90 You can try this: -stop the array -on the console/terminal type: mkdir /x mount -o skip_balance /dev/sdo1 /x Should take a few seconds to mount, once you get the cursor back type: btrfs balance cancel /x umount /x If the above commands complete normally start the array normally and post new diags.
  20. Pool is complete full and those errors are because it's running out of space, reboot to clear the read only status and you should be able to move some data elsewhere.
  21. It's not a problem with the NVMe, but it's generating correctable PCIe errors, unlikely they are related to the issues but you should still to get rid of them, if it's not possible at least suppress the errors to stop spamming the log.
  22. Unrelated to this I've been recommending for a long time monitoring any btrfs fs for errors, and it also help with this, since it's likely were any corruption issues will first be visible.
  23. If you saved the diags before rebooting please post or pm them to me if you prefer.
  24. There are some NVMe related errors, if you can can move it to a different PCIe/M.2 slot to see if they go away, but main issue are likely all the nginx errors, hopefully someone else will help with those, for now did you try booting in safe mode just to rule out any plugin issues?
  25. Your server also isn't showing any signs of corruptions so far, and it's a different model than the known affected ones, so it might be OK, again until we find out exactly what's causing this it's difficult to say if it's safe or not, best to be on the safe side for now.
×
×
  • Create New...