Jump to content

JorgeB

Moderators
  • Posts

    67,442
  • Joined

  • Last visited

  • Days Won

    706

Everything posted by JorgeB

  1. It's not a common thing but I've seen some users who usually do it like that, it's safer in case a different disk fails during the rebuild since you can use the old disk and parity will still be 100% in sync, it won't be just by starting the array normally, even if no other changes are done, and while this is usually not a big deal if you need to use the invalid slot with xfs disks it will be with btrfs disks, since they have a transaction ID and just by doing a single mount it will be out of date with the old disk.
  2. Possibly a disk issue, complete diags could give more clues, for now run an extended SMART test on that disk.
  3. This is a very weird one, are you using some kind of ECO power strip or an ECO port of an UPS? Those can sometimes be configured to cut power once it reaches a set low value, though if it was that I would expect the server to shutdown, not reboot, kind of grasping at straws here.
  4. Unrelated, those are likely board BIOS issues, though not usually important, if you have the docker image on a btrfs cache, it could be this.
  5. Not without the diags saved before rebooting, syslog is cleared after a reboot.
  6. Fine to rebuild in maintenance mode, just need to press sync to kick it off.
  7. Please post the diagnostics: Tools -> Diagnostics
  8. There are several reports in the forums of this shfs error causing /mnt/user to go away: May 14 14:06:42 Tower shfs: shfs: ../lib/fuse.c:1451: unlink_node: Assertion `node->nlookup > 1' failed. Rebooting will fix it, until it happens again, I remember seeing at least 5 or 6 different users with the same issue in the last couple of months, it was reported here that it's possibly this issue: https://github.com/libfuse/libfuse/issues/128 Attached diags from latest occurrence. tower-diagnostics-20200514-1444.zip
  9. Rebooting will fix it. User0 always exists when there's a cache device/pool.
  10. Disk looks fine, diags are after rebooting so we can't see what happened, suggest replacing/swapping cables/slot to rule them out and rebuild on top.
  11. Yep, it's a complete surface scan, the larger the disk, the longer it will take, that disk isn't looking very good, but wait for the test to finish.
  12. Extended self-test routine recommended polling time: (1382) minutes.
  13. What's the polling time for that drive?
  14. Sync errors after disk12 rebuild are likely a result of this: After that all looks normal, any more issues we need the diags when the problem happens.
  15. There were no errors on the last check so everything if fine for now, only need to worry if you get more errors on a future check (without any unclean shutdown or other issues).
  16. While we wait for the fix, anyone reached a PB? I'm not that far:
  17. Never a good sign but if the extend test is passing disks are good at least for now.
  18. Disk6 should be replaced, but lets see how disk5 does in the SMART test.
  19. SMART test passed so it's OK for now but any more errors in the near future consider replacing it.
  20. There's always some filesystem overhead.
  21. Depends on how you're doing it, but not normal, with gigabit (and mostly large files) it should be able to backup around 300GB per hour.
  22. No, v6.8.x uses a different engine, and it should mostly auto-tune, though there are some known issues for very large arrays, usually with 24+ devices.
×
×
  • Create New...