-
Posts
84 -
Joined
-
Last visited
Content Type
Profiles
Forums
Downloads
Store
Gallery
Bug Reports
Documentation
Landing
Everything posted by Tuumke
-
Yeah there is, i just can;t figure out why. It's a satisfactory container from WolveIX. I also made an issue there, as this started before my upgrade to 6.12.9.
-
I do seem to have some issues with a Satisfactory server running in a docker container. Should i be worried if my syslog is full of: Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(vethd83e8ba) entered disabled state Apr 2 07:38:52 thebox kernel: device vethd83e8ba left promiscuous mode Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(vethd83e8ba) entered disabled state Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered blocking state Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered disabled state Apr 2 07:38:52 thebox kernel: device veth0a65a1f entered promiscuous mode Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered blocking state Apr 2 07:38:52 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered forwarding state Apr 2 07:38:53 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered disabled state Apr 2 07:38:53 thebox kernel: eth0: renamed from veth1f06682 Apr 2 07:38:53 thebox kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth0a65a1f: link becomes ready Apr 2 07:38:53 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered blocking state Apr 2 07:38:53 thebox kernel: br-18371025d7b1: port 1(veth0a65a1f) entered forwarding state Apr 2 07:39:02 thebox kernel: br-f8a7f7f71814: port 22(vethe5dc7c2) entered disabled state Apr 2 07:39:02 thebox kernel: veth186df87: renamed from eth0 Apr 2 07:39:02 thebox kernel: br-f8a7f7f71814: port 22(vethe5dc7c2) entered disabled state Apr 2 07:39:02 thebox kernel: device vethe5dc7c2 left promiscuous mode Apr 2 07:39:02 thebox kernel: br-f8a7f7f71814: port 22(vethe5dc7c2) entered disabled state Apr 2 07:39:05 thebox kernel: br-f8a7f7f71814: port 22(vethaf4364a) entered blocking state Apr 2 07:39:05 thebox kernel: br-f8a7f7f71814: port 22(vethaf4364a) entered disabled state Apr 2 07:39:05 thebox kernel: device vethaf4364a entered promiscuous mode Apr 2 07:39:06 thebox kernel: eth0: renamed from veth0d85c3c Apr 2 07:39:06 thebox kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethaf4364a: link becomes ready Apr 2 07:39:06 thebox kernel: br-f8a7f7f71814: port 22(vethaf4364a) entered blocking state Apr 2 07:39:06 thebox kernel: br-f8a7f7f71814: port 22(vethaf4364a) entered forwarding state
-
Updating from 6.12.8 to 6.12.9 Stuck on "retry unmounting user shares" umount /var/lib/docker Did not work 😞 root@thebox:/mnt/user/dockers# mount rootfs on / type rootfs (rw,size=32899532k,nr_inodes=8224883,inode64) proc on /proc type proc (rw,relatime) sysfs on /sys type sysfs (rw,relatime) tmpfs on /run type tmpfs (rw,nosuid,nodev,noexec,relatime,size=32768k,mode=755,inode64) /dev/sda1 on /boot type vfat (rw,noatime,nodiratime,fmask=0177,dmask=0077,codepage=437,iocharset=iso8859-1,shortname=mixed,flush,errors=remount-ro) /boot/bzmodules on /lib type squashfs (ro,relatime,errors=continue) overlay on /lib type overlay (rw,relatime,lowerdir=/lib,upperdir=/var/local/overlay/lib,workdir=/var/local/overlay-work/lib) /boot/bzfirmware on /usr type squashfs (ro,relatime,errors=continue) overlay on /usr type overlay (rw,relatime,lowerdir=/usr,upperdir=/var/local/overlay/usr,workdir=/var/local/overlay-work/usr) devtmpfs on /dev type devtmpfs (rw,relatime,size=8192k,nr_inodes=8224883,mode=755,inode64) devpts on /dev/pts type devpts (rw,relatime,gid=5,mode=620,ptmxmode=000) tmpfs on /dev/shm type tmpfs (rw,relatime,inode64) fusectl on /sys/fs/fuse/connections type fusectl (rw,relatime) hugetlbfs on /hugetlbfs type hugetlbfs (rw,relatime,pagesize=2M) cgroup2 on /sys/fs/cgroup type cgroup2 (rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot) tmpfs on /var/log type tmpfs (rw,relatime,size=131072k,mode=755,inode64) rootfs on /mnt type rootfs (rw,size=32899532k,nr_inodes=8224883,inode64) /dev/md1p1 on /mnt/disk1 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md2p1 on /mnt/disk2 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md3p1 on /mnt/disk3 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md4p1 on /mnt/disk4 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md5p1 on /mnt/disk5 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md6p1 on /mnt/disk6 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md7p1 on /mnt/disk7 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/md8p1 on /mnt/disk8 type xfs (rw,noatime,nouuid,attr2,inode64,logbufs=8,logbsize=32k,noquota) /dev/sdj1 on /mnt/cache type btrfs (rw,noatime,discard=async,space_cache=v2,subvolid=5,subvol=/) shfs on /mnt/user type fuse.shfs (rw,nosuid,nodev,noatime,user_id=0,group_id=0,default_permissions,allow_other) tmpfs on /run/user/0 type tmpfs (rw,nosuid,nodev,relatime,size=6583272k,nr_inodes=1645818,mode=700,inode64) -edit- Rebooted manually. Stopped array again, worked this time. Updating went without any issues after that
-
I know have errors in my docker-compose.yml? line 77: mapping key "x-environment" already defined at line 72 line 83: mapping key "x-environment" already defined at line 72 line 83: mapping key "x-environment" already defined at line 77 71 # Common environment values 72 x-environment: &default-tz-puid-pgid 73 TZ: $TZ 74 PUID: $PUID 75 PGID: $PGID 76 77 x-environment: &projectsend-tz-puid-pgid 78 TZ: $TZ 79 PUID: $PUID 80 PGID: $PGID 81 MAX_UPLOAD: 5000 82 83 x-environment: &ls-dockermods-ping 84 DOCKER_MODS: linuxserver/mods:universal-package-install 85 INSTALL_PACKAGES: iputils-ping How do i go back to newer docker? I had 0 problems with the new one (maybe because i have 512gb memory) -edit- Looks like i need a newer docker-compose? -edit2- Think i didn't fix this before upgrading. Seems you can't use multiple x-environments but need to make them unique. Never looked into this before my 5 day leave 😛
-
I haven't updated the parity check tuning plugin (still on 2022.12.05) and also have the issue of not being able to login anymore -edit- Also, when trying to update that plugin: plugin: updating: parity.check.tuning.plg plugin: downloading: parity.check.tuning-2023.02.10.txz ... plugin: parity.check.tuning-2023.02.10.txz download failure: Invalid URL / Server error response Executing hook script: post_plugin_checks -edit2- a reboot of the system resolved it for me (not having updated the plugin). Don't think it's related to the plugin -edit-3 Come to think of it, i think i was trying to SSH into my VM which wasn't powered on... So not 100% sure with the above
-
Anyone? 😮
-
Hey guys, I got my dedicated machine at Hetzner up and running for a while now. Thought i'd play with a VM for the first time on Unraid. As soon as i launched it, i got an e-mail from hetzner support saying they see a macadress on my public IP that doesn't belong there. So shut it down and deleted it again. After doing some research, i understand that i needed an extra public IP address, linked to a provided (by hetzner) macaddress. So i did, bought a mac+pubip with hetzner, assigned it to a VM and it now has it's own public IP address. The strange thing is, that i cannot access it from the outside, just the VNC console. There is no firewall active on the hetzner dasbhoard, i made sure it's disabled. Are there any things i need to do on Unraid to make it possible to VNC or even SSH into the VM?
-
Just updated from 6.11.1 to 6.11.5. Will post if anything is wrong
-
@jbreed you still maintaining this docker?
-
Guess that makes sense. It's the Accumilation duration? That's set to 1 hour: This is a "new" setup. So all parity should be pretty default?
-
Thanks for your reply. I though i had deinstalled nerdpack! Thanks for noticing. I did sometimes need to manually unrar files, so i used unrar. But have that installed now via a package manually as explained in the threads. What things could cause the parity check to pause? -edit- My parity check was planned on Sundays, 02:00 and Mover daily on 03:40. Think that might have been it? Changed mover to 01:00. See if that works
-
I've attached a diagnostic package. It keeps happening each time! I just don't understand why. thebox-diagnostics-20221030-1417.zip
-
Seems it wouldn't do this if i selected GPU to me my onboard card? Now set to Virtual and i got the KVM/VNC options.
-
-
How do you KVM to a new VM? I tried to create a VM and it was started, but i had no option to open a console or anything.
-
And why is that necessary?
-
I had setup my new dedicated (hetzner) machine to use 2 parity disks. Is it possible to go back to 1 parity disk? Or will that break my system
-
I have not tried that latest release on 6.11.0. Have you updated unRAID yet? I have but not the docker container I have too much in que still since this is a new server
-
unrar is one i found missing now -edit- Fixed it: ssh session on Unraid: cd /boot mkdir extra cd extra wget https://slack.conraid.net/repository/slackware64-current/unrar/unrar-6.1.7-x86_64-1cf.txz pkginstall unrar-6.1.7-x86_64-1cf.txz
-
I didn't know. Now what? How do i get this back to working?
-
I had issues with SAB vpn connection. The startup of the container is so super slow! Switched to binhex/arch-sabnzbdvpn:3.5.3-1-03 and that worked like a charm? It also made my system feel super sluggish.
-
unRAID 6 NerdPack - CLI tools (iftop, iotop, screen, kbd, etc.)
Tuumke replied to jonp's topic in Plugin Support
That fixed it! Thank you so much -
unRAID 6 NerdPack - CLI tools (iftop, iotop, screen, kbd, etc.)
Tuumke replied to jonp's topic in Plugin Support
I hadn't, no. I'm assuming top directory is / ? root@thebox:/# iotop.py -bash: iotop.py: command not found root@thebox:/# -
unRAID 6 NerdPack - CLI tools (iftop, iotop, screen, kbd, etc.)
Tuumke replied to jonp's topic in Plugin Support
root@thebox:/mnt/user/system/docker# ln /usr/bin/python3 /usr/bin/python root@thebox:/mnt/user/system/docker# python --version Python 3.9.6 root@thebox:/mnt/user/system/docker# iotop No module named 'iotop' To run an uninstalled copy of iotop, launch iotop.py in the top directory root@thebox:/mnt/user/system/docker# Tried reinstalling iotop, didnt work -
unRAID 6 NerdPack - CLI tools (iftop, iotop, screen, kbd, etc.)
Tuumke replied to jonp's topic in Plugin Support
Any idea how to fix iotop? root@thebox:/# iotop -bash: /usr/sbin/iotop: /usr/bin/python: bad interpreter: No such file or directory root@thebox:/# python3 --version Python 3.9.6