• Content Count

  • Joined

  • Last visited

  • Days Won


Everything posted by JorgeB

  1. JorgeB


    Please post the diagnostics: Tools -> Diagnostics
  2. I would, since there are several bug fixes and it's very easy to do.
  3. It is in IT mode, but latest firmware is
  4. These are SAS only for now and the foreseeable future, also, and as far as I understand, they are presented to the OS as two 7TB drives, using dual LUNs, so no performance gains with Unraid and if one disk failed you'd lose two assigned drives.
  5. Disk dropped offline, change/replace cables to rule them out see if it happens again to the same disk, disk will need to be rebuilt.
  6. You can still use the new alignment with v6.8 if you use a cache pool, a single device cache can be a "pool" as long as cache slots are set to more than 1.
  7. There are multiple duplicate UUIDs, please post output of blkid
  8. You can do a new config if you want to do it all at the same time, array will be unprotected until the sync finishes.
  9. I would try a different power supply and/or connecting the Seagates to a different controller, like the onboard SATA ports.
  10. It's up to you, btrfs works great with good hardware, and unlike xfs it will easily detect data corruption, but if you don't need any of its features and plan to use a single device you can switch to xfs. Data will remain corrupt, and possibly the filesystem, might be better to recreate.
  11. That rules out the network, or at least it's not the only problem, have you tried copying large files to the array with turbo write enable?
  12. And btrfs is much more sensitive to RAM errors than for example xfs, so much more likely for problems to be detected there first, and a couple of bit flips in the wrong place can destroy a btrfs filesystem.
  13. Yep, try with just one dimm to find the problem one.
  14. Feb 26 09:58:52 Trashnet kernel: BTRFS info (device nvme0n1p1): bdev /dev/nvme0n1p1 errs: wr 0, rd 0, flush 0, corrupt 12722, gen 0 Feb 26 09:58:52 Trashnet kernel: BTRFS info (device nvme0n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 0, flush 0, corrupt 17619, gen 0 Both devices are detecting a lot of data corruption, start by running memtest.
  16. I have two of these: You could install 6 disks on the 4 bays, but got them a long time ago, so probably discontinued, though there should be something similar.
  17. Nothing else, but one thing I forgot to mention is before doing that make sure cache backups are up to date, just in case.
  18. To reboot you don't need to power cycle, power cycling is shutting down the server, wait a few seconds, power back on.
  19. Just the power cycle will likely bring the disk back (not a reboot), but might as well check/replace cables to rule them out.
  20. Syslog is showing ATA9 errors, ATA9 is disk5, if errors persist after changing cables post a SMART report for that disk.
  21. And the iperf test was single stream like mentioned? If yes see how the server performs with internal transfers, best to disable parity first if not that can be the bottleneck.
  22. Disk dropped offline, check connections and the disk should come back up after a power cycle, then run the SMART test.
  23. I believe there have been other reports of issues with those disks when used on a LSI with v6.9, possibly a driver issue, you could try connecting them to the onboard SATA, of course swap with disks from a different model.
  24. You don't need that with updated firmware.
  25. Are both disks from the same model?