WEHA

Members
  • Posts

    83
  • Joined

  • Last visited

Everything posted by WEHA

  1. I had a share that was set to cache prefer on cache nr 2. I want to get rid of cache 2 to replace hdd's with ssd's. So I changed the setting to cache yes on cache nr 1. When I started mover it wanted to move the files on cache nr 2 but it said "file exists": move: move_object: /mnt/cache2/xxx.yyy File exists When I set it to cache yes on cache nr 2 and restarted mover, it started working again.
  2. That should not be necessary at all, exceptions is when all else fails. Anyway, it does not really matter, it's "fixed" in beta35
  3. Very well, thanks for your input.
  4. Well yes, not via btrfs but I have no issues with the vm, no errors in eventlog and full backups are working. That's why I believe the vdisk is fine. It's just weird to me that only docker image is affected and it was on a COW share. But if you're confident that there is no issue with this scenario then ok.
  5. I mean COW by enabling. So system had COW and the vdisk had NOCOW But docker image was corrupt and vdisk image was not.
  6. It's strange that it's only the docker file and not the vm file... could it be related to NOCOW / COW? I enabled this for the system share and thus the docker image, the vdisk has NOCOW. Thank you for assisting
  7. I moved everything off, 2 files remained, 1 vdisk file and docker img. The docker image was unable to be moved due to an i/o error, so I removed it and recreated it on another pool. I reran scrub and now no errors are detected. Is this related to docker image being set as xfs on a btrfs pool? I set this to xfs to be sure the bug that causes much disk i/o to be gone. Smart does not show any errors on the disk so I can be sure this was a software corruption and not caused by a hardware (hdd) defect?
  8. Same story, I see callbacks suppressed though [203355.213783] BTRFS error (device sde1): unable to fixup (regular) error at logical 1342354677760 on dev /dev/sde1 [203436.360164] scrub_handle_errored_block: 8 callbacks suppressed [203436.360209] btrfs_dev_stat_print_on_error: 8 callbacks suppressed [203436.360212] BTRFS error (device sde1): bdev /dev/sde1 errs: wr 0, rd 0, flush 0, corrupt 93, gen 0 [203436.360214] scrub_handle_errored_block: 8 callbacks suppressed [203436.360215] BTRFS error (device sde1): unable to fixup (regular) error at logical 1348826648576 on dev /dev/sde1 [203439.353192] BTRFS error (device sde1): bdev /dev/sde1 errs: wr 0, rd 0, flush 0, corrupt 94, gen 0 [203439.353195] BTRFS error (device sde1): unable to fixup (regular) error at logical 1349298642944 on dev /dev/sde1 [203440.426170] BTRFS error (device sde1): bdev /dev/sde1 errs: wr 0, rd 0, flush 0, corrupt 95, gen 0 [203440.426174] BTRFS error (device sde1): unable to fixup (regular) error at logical 1349556105216 on dev /dev/sde1 [203441.204687] BTRFS error (device sde1): bdev /dev/sde1 errs: wr 0, rd 0, flush 0, corrupt 96, gen 0 [203441.204690] BTRFS error (device sde1): unable to fixup (regular) error at logical 1349681184768 on dev /dev/sde1
  9. tower-diagnostics-20201124-1452.zip
  10. It's copied from the syslog file in nano, so I would think that is the full syslog? There are warnings from before the scrub though: root@Tower:/var/log# cat syslog |grep "BTRFS warning" Nov 23 03:59:25 Tower kernel: BTRFS warning (device sde1): csum failed root 5 ino 182291 off 1765621760 csum 0xd488241c expected csum 0xdbe78a4e mirror 1 Nov 23 03:59:25 Tower kernel: BTRFS warning (device sde1): csum failed root 5 ino 182291 off 1765621760 csum 0xd488241c expected csum 0xdbe78a4e mirror 1 Nov 23 20:40:23 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 281 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 Nov 23 20:40:23 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 281 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 Nov 24 04:03:17 Tower kernel: BTRFS warning (device sde1): csum failed root 5 ino 182291 off 4379881472 csum 0x1616fb61 expected csum 0xcbd3dbb1 mirror 2 Nov 24 09:19:19 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 282 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 Nov 24 09:19:19 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 282 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 Nov 24 09:22:05 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 283 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 Nov 24 09:22:06 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 283 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1
  11. Syslog does not show files: Nov 24 13:01:51 Tower kernel: BTRFS info (device sde1): scrub: started on devid 1 Nov 24 13:01:51 Tower kernel: BTRFS info (device sde1): scrub: started on devid 2 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 1, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 2, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 3, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 4, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 6, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 6, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 7, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 8, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413978710016 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 9, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413913239552 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413913341952 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413913444352 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 10, gen 0 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413915201536 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413915303936 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413915406336 on dev /dev/sdk1 Nov 24 13:03:22 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413916004352 on dev /dev/sdk1 Nov 24 13:03:23 Tower kernel: BTRFS error (device sde1): fixed up error at logical 1413978824704 on dev /dev/sdk1 Nov 24 13:03:23 Tower kernel: BTRFS error (device sde1): unable to fixup (regular) error at logical 1413979930624 on dev /dev/sdk1
  12. *sigh* ... how do I get a list of files? I'm running scrub and this is the status already: Error summary: csum=35 Corrected: 4 Uncorrectable: 31 Unverified: 0 These are software errors, correct? Smart does not indicate a problem, this is also a new disk.
  13. Attached Seems like this is the curlprit? Nov 24 09:22:05 Tower kernel: BTRFS warning (device sde1): csum failed root -9 ino 283 off 951992320 csum 0x47d58bec expected csum 0x56997f79 mirror 1 tower-diagnostics-20201124-1149.zip
  14. Tried converting twice, remains the same state as posted earlier. It starts and after about 30 seconds or so it goes back to no balance.
  15. Could you just confirm to me if converting from single to raid 1 does not lose data? (not stated in faq nor unraid gui) I just added a disk to a cache pool from 1 to 2 and unraid made it single. (I believe this is the default according to the faq) So this is the current state (2 states, related to the btrfs bug?): Data, RAID1: total=42.00GiB, used=24.68GiB Data, single: total=1.18TiB, used=1.16TiB System, DUP: total=8.00MiB, used=176.00KiB Metadata, DUP: total=2.00GiB, used=1.69GiB GlobalReserve, single: total=512.00MiB, used=0.00B I have enough space available so nothing will happen to my data right? What would happen if there was not enough space?
  16. I'm trying to create a JBOD cache pool in 6.9beta35. I don't know if this is a bug or I'm just doing it wrong so... From what I understand from the below post I have to set it to single mode When I do this "convert to single mode" (it's a 14TB and 8TB disk) the GUI says it's 16TB. I also see the same write speeds to both disks, giving the impression it's RAID 1 Balance status: Data, RAID1: total=4.00GiB, used=2.97GiB Data, single: total=1.00GiB, used=0.00B System, RAID1: total=32.00MiB, used=16.00KiB Metadata, RAID1: total=1.00GiB, used=3.94MiB GlobalReserve, single: total=3.78MiB, used=16.00KiB If I execute "perform full balance", it just reverts to RAID 1 status. Can anyone tell me what I'm doing wrong or should I post this as a bug in beta? Maybe I have to jump through a few hoops like removing one disk -> single mode -> add disk? thanks!
  17. I have 2 linux vm's, ubuntu 18 & centos 8, 2 x freebsd and 4 windows. Only the linux vm's loose network over time. Restarting the network on the vm itself via vnc makes it work again (at least on the ubuntu), after a vm reboot it also works again. Not sure what to provide for extra information, let me know what you want. There is nothing special I can see in the syslog nor vm log, still have to check the linux log itself. I've changed the network adapter to virtio-net to see if that makes a difference. Can anyone tell me what the difference is between virtio & virtio-net?
  18. I'm not sure if this is a problem in the gui or a trigger happy antivirus but I thought I'd mention it: Bitdefender Endpoint Security Tools blocked this page The page you are trying to access contains a malicious attack attempt. Detected attack: Exploit.CommandInjection.Gen.123 Access from your browser has been blocked.
  19. So I found some threads going back years with the same endpoint error. After a reboot, all shares returned to normal. Still trying to recover from the "share is deleted message" I hope you can find the curlprit with my logs. I found this: Nov 8 18:04:32 Tower emhttpd: Starting services... Nov 8 18:04:32 Tower move: move: file .... Nov 8 18:04:32 Tower kernel: shfs[94877]: segfault at 0 ip 000000000040546a sp 000014710c7c7840 error 4 in shfs[403000+d000] Nov 8 18:04:32 Tower kernel: Code: 48 8b 45 f0 c9 c3 55 48 89 e5 48 83 ec 20 48 89 7d e8 48 89 75 e0 c7 45 fc 00 00 00 00 8b 45 fc 48 63 d0 48 8b 45 e0 48 01 d0 <0f> b6 00 3c 2f 74 43 8b 0> Nov 8 18:04:32 Tower move: move: create_parent: /mnt/disk6 error: Software caused connection abort
  20. I was moving data from the array to the cache with mover. I noticed that the share was going to be too big to fit. I then changed the share setting to "no" and when I clicked apply it said "share sync has been deleted" After having a small heartattack I checked the disks & cache folder for the share, it was still there. However /mnt/user now gives this error: -bash: cd: user: Transport endpoint is not connected Clicking "shares" in the gui only shows an entry in disk shares called cache. I assume I can just stop and start the array to get everything working again? My vm's are still running... tower-diagnostics-20201108-1812.zip
  21. When making some changes it's sometimes necessary or preferred to start docker / vm manager without auto-start enabled. That way you can just start whichever docker or vm you want. So what I'm asking is: add a 3rd option to the enable docker / enable vms dropdown like Yes, no auto start
  22. So I read about the "bug" that causes many writes to sdd's especially evo... Mine have a 1200TBW and are around 1500TBW now (in 2 years time) In the new beta there is a solution, but there also issues. My thought is, can I upgrade to the new beta, recreate the cache (on new drives) with the new partition layout and revert back to 6.8.3 if the need arises?