July 2, 201510 yr Nah, just the appdata share since it seems to be the template default. Squid is already on it with the latest version of Community Applications plugin. Experimental (though it shouldn't cause any issues anywhere) and optional.
July 3, 201510 yr Author I updated my UNRAID version to 6.0.1 in the hope that would eventually fix any Docker issues (even though there is nothing in the release notes about this) but the problem persists. I can reproduce a "crash" of my containers and provide any other logs and diagnostics if this would help?
July 4, 201510 yr Author Hopefully not too off topic as it is related to my cache disk and therefore potentially my docker image. I wanted to replace my cache disk to eliminate this as the source of error, the only files that I have there that I want to keep are my VMs. One VM copied fine but I'm getting errors during the copy of the other VM (cp: error reading ‘vdisk1.img’: Input/output error) I checked the destination and specified the disk during the copy and there is plenty of room. Therefore, does anyone have am idea of how I can force the copy? Could this indicate an error with the cache disk? I ran a SMART disk check and this came up clean. Thanks
July 4, 201510 yr For those having no docker tab, did you check the following: Do not have the old Dockerman plugin Do not have a corrupted or full docker.img file Do not have a full log system For the last 2 points use the following command df -h Then look at /var/log and /var/lib/docker (i think thats the locations)
July 4, 201510 yr Author After a reboot the docker tab came back, however the problems with containers crashing persist. I installed v6.0.0 cleanly, formatting the flash drive beforehand, so there were no plugins. I've recreated the docker.img a few times, however I'm sure you've seen my previous posts explaining, docker runs fine for a time upon rebooting then the issues come back. How do I check if the system logs are full? There's plenty of room on the flash drive. df -h gives this, Filesystem Size Used Avail Use% Mounted on tmpfs 128M 40M 89M 31% /var/log /dev/sda1 3.8G 290M 3.5G 8% /boot /dev/md1 1.9T 1.9T 696M 100% /mnt/disk1 /dev/md2 1.9T 1.9T 13G 100% /mnt/disk2 /dev/md3 1.9T 1.5T 407G 79% /mnt/disk3 /dev/md4 3.7T 3.2T 466G 88% /mnt/disk4 /dev/md5 1.9T 1.8T 21G 99% /mnt/disk5 /dev/md6 1.9T 1.9T 14G 100% /mnt/disk6 /dev/md7 1.9T 1.8T 42G 98% /mnt/disk7 /dev/md8 1.9T 1.8T 84G 96% /mnt/disk8 /dev/md9 1.9T 1.8T 54G 98% /mnt/disk9 /dev/md10 1.9T 1.9T 1.7G 100% /mnt/disk10 /dev/md11 1.9T 1.8T 30G 99% /mnt/disk11 /dev/md12 1.9T 1.8T 54G 98% /mnt/disk12 /dev/md13 1.9T 1.4T 463G 76% /mnt/disk13 /dev/md14 1.9T 934G 930G 51% /mnt/disk14 /dev/md15 1.9T 1.1T 754G 60% /mnt/disk15 /dev/sdn1 466G 180G 286G 39% /mnt/cache shfs 30T 26T 3.3T 89% /mnt/user0 shfs 30T 27T 3.6T 89% /mnt/user /dev/loop1 1.8M 89K 1.6M 6% /etc/libvirt /dev/loop0 15G 2.3G 12G 17% /var/lib/docker /var/log/syslog gives this, (the last few lines) Jul 4 15:19:32 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:19:32 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:19:32 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:30:36 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:30:36 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:30:36 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:31:41 msstorage kernel: mdcmd (353): spindown 3 Jul 4 15:31:49 msstorage kernel: mdcmd (354): spindown 9 Jul 4 15:31:57 msstorage kernel: mdcmd (355): spindown 7 Jul 4 15:32:04 msstorage kernel: mdcmd (356): spindown 6 Jul 4 15:32:04 msstorage kernel: mdcmd (357): spindown 10 Jul 4 15:32:22 msstorage kernel: mdcmd (358): spindown 5 Jul 4 15:32:22 msstorage kernel: mdcmd (359): spindown 11 Jul 4 15:34:01 msstorage kernel: mdcmd (360): spindown 1 Jul 4 15:34:01 msstorage kernel: mdcmd (361): spindown 4 Jul 4 15:39:10 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:39:10 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:39:10 msstorage kernel: BTRFS warning (device sdn1): csum failed ino 273 off 7155044352 csum 2214257214 expected csum 3520190875 Jul 4 15:40:55 msstorage kernel: verify_parent_transid: 2 callbacks suppressed Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127549440 wanted 2290 found 1968 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127549440 wanted 2290 found 1968 Jul 4 15:40:55 msstorage kernel: BTRFS info (device loop0): no csum found for inode 264 start 0 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127549440 wanted 2290 found 1968 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127549440 wanted 2290 found 1968 Jul 4 15:40:55 msstorage kernel: BTRFS info (device loop0): no csum found for inode 264 start 4096 Jul 4 15:40:55 msstorage kernel: BTRFS warning (device loop0): csum failed ino 264 off 0 csum 3551632587 expected csum 0 Jul 4 15:40:55 msstorage kernel: BTRFS warning (device loop0): csum failed ino 264 off 4096 csum 3166042738 expected csum 0 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 126992384 wanted 2290 found 2289 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 126992384 wanted 2290 found 2289 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 126992384 wanted 2290 found 2289 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 126992384 wanted 2290 found 2289 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127926272 wanted 2290 found 1969 Jul 4 15:40:55 msstorage kernel: BTRFS (device loop0): parent transid verify failed on 127926272 wanted 2290 found 1969 Jul 4 15:40:57 msstorage kernel: mdcmd (362): spindown 2 Jul 4 15:40:58 msstorage kernel: mdcmd (363): spindown 14 Jul 4 16:00:39 msstorage kernel: mdcmd (364): spindown 13 /var/lib/docker, hmmm this is interesting, I tried to list the files here and have the same error as when I try to copy from the cache disk. root@msstorage:/var/lib/docker# ls /bin/ls: reading directory .: Input/output error
July 4, 201510 yr Whatever system you are using BTRFS on is corrupted. One such system everyone has is the docker.img is internally a BTRFS filesystem. You can try running a scrub on it to fix it, but i have never seen anyone have any luck on uncorrupting it other than having to delete it and recreate it. Some say BTRFS is stable and mature, but i can't think it is when there are so many corruptions reported that are unfixable.
July 4, 201510 yr Author Thanks BRiT. I don't care about the docker.img but I would really like my VM back if possible. I'm running the scrub now and post the results once finished.
July 4, 201510 yr Author Does anyone run docker on a non BTRFS file system? What are your experiences? I believe I remember reading that this file system was better to compression or maybe de-duplication but I'm willing to sacrifice that on a more stable file system for docker.
July 4, 201510 yr Does anyone run docker on a non BTRFS file system? What are your experiences? I believe I remember reading that this file system was better to compression or maybe de-duplication but I'm willing to sacrifice that on a more stable file system for docker. On unRAID it's not possible without great trouble to run docker on non BTRFS since LimeTech's implementation uses BTRFS inside the "docker.img" file. Now if you meant what filesystem is used to store the docker.img, then I have that on a XFS cache drive.
July 4, 201510 yr Author Ah ok, If I understand then the filesystem inside the docker file is BTRFS and the docker image can be stored on any file system? Thanks for the info.
July 4, 201510 yr Ah ok, If I understand then the filesystem inside the docker file is BTRFS and the docker image can be stored on any file system? Thanks for the info. Correct.
July 4, 201510 yr I'm wondering if all of the host paths are set up correctly? Could improper host paths where the containers are storing their metadata within docker.img be causing the image to fill up and corrupt?
July 4, 201510 yr Author The path I use for container configuration is /mnt/disk15/Docker-Configs. Is this wrong? Here is the result from the read only scrub (This is from the disk management page and not docker) scrub started at Sat Jul 4 16:57:41 2015, running for 2544 seconds total bytes scrubbed: 179.39GiB with 13 errors error details: csum=13 corrected errors: 0, uncorrectable errors: 0, unverified errors: 0 Would you advise that I redo this and try to correct errors? Would this in any way create more of a risk to my VM that I can't copy? Thanks
July 4, 201510 yr Author Whatever system you are using BTRFS on is corrupted. One such system everyone has is the docker.img is internally a BTRFS filesystem. You can try running a scrub on it to fix it, but i have never seen anyone have any luck on uncorrupting it other than having to delete it and recreate it. Some say BTRFS is stable and mature, but i can't think it is when there are so many corruptions reported that are unfixable. I guess you were right :-( scrub started at Sat Jul 4 19:17:11 2015 and finished after 2531 seconds total bytes scrubbed: 179.49GiB with 13 errors error details: csum=13 corrected errors: 0, uncorrectable errors: 13, unverified errors: 0
July 8, 201510 yr Author So after many hours of trying different methods of copying the VM I wanted off my cache drive I gave up after all the errors. I could however still run the VM without any visible issues and I backed up the server from within the guest OS. I changed my cache drive to be sure (although I believe the error to have been on the BTRFS file system and not the disk) as I had a spare 500GB disk doing nothing. I formated this to XFS and re-did all my VMs and docker containers, 2 days running and so far so good. Thanks for the advice given here and I won't be using BTRFS on my cache drive any time soon
Archived
This topic is now archived and is closed to further replies.