Ryonez

Members
  • Posts

    159
  • Joined

  • Last visited

About Ryonez

  • Birthday April 19

Converted

  • Gender
    Male

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

Ryonez's Achievements

Apprentice

Apprentice (3/14)

22

Reputation

  1. Issue is present in version 6.12.0 as well. It is getting very tedious to baby docker along... I can't even do backup because it'll trigger this. Here's the latest diagnostics. atlantis-diagnostics-20230616-0553.zip
  2. So, for some behavioural updates. Noticed my RAM usage was at 68, not the 60 it normally rests at. Server isn't really doing anything, so figured it was from the two new containers, diskover and elasticsearch, mostly from elasticsearch. So I shut that stack down. That caused docker to spike the CPU and RAM a bit. The containers turned off in about 3ish seconds, and RAM usage dropped to 60. Then, docker started doing it's thing: Ram usage started spiking up to 67 in this image, and CPU usage is spiking as well. After docker settles with whatever it's doing: This is from just two containers being turned off. This gets worse the more docker operations you preform.
  3. Diskover just happened to be what I was trying out yesterday when the symptoms occurred yesterday. It has not been present during the earlier times the symptoms occurred. After rebuilding docker yesterday I got diskover working and had it index the array, with minimal memory usage (I don't think I saw diskover go above 50mb). I'm not really sure how this is meant to help with finding folders using memory. It actually suggesting it's moving log folders to ram, which would increase the ram usage. If the logs are causing a blowout of data usage, that usage would go to the cache/docker location, not memory. It might be good to note I use docker in the directory configuration, not the .img, so I am able to access those folders if needed.
  4. Just a heads up on my progress from this morning. My containers are back up, didn't pass 50% of my RAM during this. So currently up and running until docker craps itself again/
  5. I don't believe this to be an issue with the docker containers memory usage. For example, the three images I was working on today: scr.io/linuxserver/diskover docker.elastic.co/elasticsearch/elasticsearch:7.10.2 alpine (elasticsearch-helper) diskover wasn't able to connect to elasticsearch, elasticsearch was crashing due to perm issues I was looking at, and elasticsearch-helper runs a line of code and shuts down. I don't see any of these using 10+GB of ram in a non functional state. The system when running steady uses 60% of 25GB of ram. And this wasn't an issue with the set of containers I was using until the start of this month. I believe this to be an issue in docker itself currently.
  6. When working with docker, I've been having issues for the past month, though not without finding what exactly the trigger is. The first time this happened was when I update all containers, around the 1st of May for me. Forced an unclean reboot after oom-killer was triggered, and the containers loaded up all right after the OS came back up, albeit with high CPU usage, though I can't remember if the memory usage spiked at the time. The second was a few days later, May the 4th for me. Again triggered during an update, forced another unclean reboot after oom-killer was triggered, this time the containers did not come back up, docker chocked the system up both on CPU and memory. Symptoms: Changes to the docker image seem to illicit high CPU usage from docker. Is also causes high spikes in memory usage. The spike in memory usage can memory to get completely used, triggering the oom-killer. The more containers in use, the more likely it seems to cause this every time. The amount required to trigger oom-killer is unknown currently. Things tried to resolve this: Checking the docker image filesystem (Result: No problems). Deleting and rebuilding the docker image, several times. Did a cache filesystem check (BTRFS, Result: Errors found.) Replaced the cache drives completely. Re-flashing the the boot USB and only copying configs back over. Starting shifting almost all my docker container configs to portainer, as docker template duplicate fields and restores unwanted fields from template updates. "Currently" (I'm having to rebuild the docker image again atm), only 3-4 containers are unRaid managed. Currently it seems if I only load containers up in small batches, I can get every container up. But only from a clean docker image, or when editing a small number of containers at a time (this I'm feeling iffy on). For the first diagnostic log attached (atlantis-diagnostics-20230507-1710), this seems to have logs from the 4th of May mentioned above. It'll have the logs from when docker was no longer able to be started. I ended up leaving docker off when it triggered the oom-killer. the log on the 7th shows that zombie process from the docker containers where preventing unmounting the cache. and unclean reboot was done. For the second diagnostic log attached (atlantis-diagnostics-20230521-1029), I was working on getting a small container stack of 3 images going. But I was configuring them, meaning they were reloaded several times in a row, and seemly creating a situation where the CPU and Memory spikes got stacked up enough for docker to trigger the oom-killer. Marking this as urgent as docker killing itself via oom-killer is leaving a big risk of data corruption, along with leaving docker pretty much useless currently. atlantis-diagnostics-20230507-1710.zip atlantis-diagnostics-20230521-1029.zip
  7. I've been experiencing what appears to be this issue. I've been shifting almost all of my docker management over to Portainer. My best guess so far is that the templates added fields back into templates that started screwing with the docker image and broke things somehow. I've been trying pretty much everything, deleting the docker img, even replaced the cache drives, reformatting the os drive. The only things kept that entire time was the config files. There's still some dockers to shift over, but most are on Portainer, and things seem to be steady. This issue first appeared when updating some images. Otherwise, I'd been running what I had for over a year.
  8. Changing the port was not done to secure it. It was done so a docker container could use it. Only specific ports are exposed, wireguard is used to connect when out of network for management, as management ports are not meant to be exposed. The issue is unRaid starts it with default configs, creating a window of attack until it restarts services with the correct configs. This is not expected behaviour. Diagnostics have been added to the main post.
  9. Hi there. I had to reboot my server recently, and I received a `Possible Hack Attempt` notification from a plugin called `Fix Common Problems'. Looking into this, I think I've found a bug/security issue. The server's SSH port should be configured to port `1122`. However while the system is booting up, SSH is started before the set configuration is read, leading to port 22 being exposed. Now, I have port 22 forwarded to the server for a service running `Forgejo`, a code repository service. The reason port 22 is being used is to lessen client configuration, and to make it simpler for users. The bug here is unRaid is loading SSH with default values for a time, and later restarting with the correct config. This creates a window for attackers to attempt to access the system through something that isn't meant to be exposed. Steps to reproduce: 1. Set a different port for SSH through unjRaid's web UI. 2. Reboot the server. 3. There is now a windows of opportunity to attack the server via the default SSH port as it starts. Plugins extend this window. Some of the Logs: SSH being started with port 22: SSH login attempts being made during this window of opportunity: There's more than just those. SSH being restarted with the correct configuration: I'll also make a quick note it looks like smbd is having the same treatment, it's loaded with defaults, then restarted with the proper configs later. So it appears there might be several system services being started like this when they shouldn't. Will make this as urgent, though I'll admit I'm not 100% if SSH will accept the correct password. This server runs a chat service, so uptime is import for me, and testing would interrupt that. Please adjust the urgency as needed if it's not as bad as I think it might be. atlantis-diagnostics-20230227-0254.zip
  10. Been working with a VaultHunters3 server using your image, and I've noticed the the minecraft servers aren't being cleanly shutdown when the container is shutdown. Is there a way to fix this? Clean shutdowns are very much preferred.
  11. Second week in a row: Apr 25 02:10:37 Atlantis CA Backup/Restore: docker stop -t 60 atlantis-yoko Apr 25 02:10:37 Atlantis CA Backup/Restore: Backing up USB Flash drive config folder to Apr 25 02:10:37 Atlantis CA Backup/Restore: Using command: /usr/bin/rsync -avXHq --delete --log-file="/var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log" /boot/ "/mnt/user/Backups/machine_specific_backups/atlantis_flash/" > /dev/null 2>&1 Apr 25 02:10:42 Atlantis CA Backup/Restore: Changing permissions on backup Apr 25 02:10:42 Atlantis CA Backup/Restore: Backing up libvirt.img to /mnt/user/Backups/machine_specific_backups/atlantis_libvirt/ Apr 25 02:10:42 Atlantis CA Backup/Restore: Using Command: /usr/bin/rsync -avXHq --delete --log-file="/var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log" "/mnt/user/system/libvirt/libvirt.img" "/mnt/user/Backups/machine_specific_backups/atlantis_libvirt/" > /dev/null 2>&1 Apr 25 02:10:42 Atlantis CA Backup/Restore: Changing permissions on backup Apr 25 02:10:42 Atlantis CA Backup/Restore: Backing Up appData from /mnt/user/appdata/ to /mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected] Apr 25 02:10:42 Atlantis CA Backup/Restore: Using command: cd '/mnt/user/appdata/' && /usr/bin/tar -cvaf '/mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected]/CA_backup.tar' --exclude "alteria-plex/Library/Application Support/Plex Media Server/Metadata" --exclude "0-pihole" --exclude "alteria-postgres" --exclude "alteria-hydrus-server" * >> /var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log 2>&1 & echo $! > /tmp/ca.backup2/tempFiles/backupInProgress Apr 25 02:10:42 Atlantis CA Backup/Restore: Backup Complete Apr 25 02:10:42 Atlantis CA Backup/Restore: Verifying backup Apr 25 02:10:42 Atlantis CA Backup/Restore: Using command: cd '/mnt/user/appdata/' && /usr/bin/tar --diff -C '/mnt/user/appdata/' -af '/mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected]/CA_backup.tar' > /var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log & echo $! > /tmp/ca.backup2/tempFiles/verifyInProgress Apr 25 02:10:42 Atlantis kernel: br-55ef392db062: port 1(veth391173c) entered blocking state Apr 25 02:11:43 Atlantis kernel: br-55ef392db062: port 70(vetha9a6916) entered forwarding state Apr 25 02:11:44 Atlantis CA Backup/Restore: ####################### Apr 25 02:11:44 Atlantis CA Backup/Restore: appData Backup complete Apr 25 02:11:44 Atlantis CA Backup/Restore: ####################### Apr 25 02:11:44 Atlantis CA Backup/Restore: Deleting /mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected] Apr 25 02:11:45 Atlantis CA Backup/Restore: Backup / Restore Completed That's two weeks of missed backups, this is a big issue.
  12. Just had the plugin fail to backup. Looking at the logs, everything worked fine, verification was good, but the backup file just does not exist. Apr 18 02:10:41 Atlantis CA Backup/Restore: docker stop -t 60 atlantis-yoko Apr 18 02:10:41 Atlantis CA Backup/Restore: Backing up USB Flash drive config folder to Apr 18 02:10:41 Atlantis CA Backup/Restore: Using command: /usr/bin/rsync -avXHq --delete --log-file="/var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log" /boot/ "/mnt/user/Backups/machine_specific_backups/atlantis_flash/" > /dev/null 2>&1 Apr 18 02:10:44 Atlantis CA Backup/Restore: Changing permissions on backup Apr 18 02:10:44 Atlantis CA Backup/Restore: Backing up libvirt.img to /mnt/user/Backups/machine_specific_backups/atlantis_libvirt/ Apr 18 02:10:44 Atlantis CA Backup/Restore: Using Command: /usr/bin/rsync -avXHq --delete --log-file="/var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log" "/mnt/user/system/libvirt/libvirt.img" "/mnt/user/Backups/machine_specific_backups/atlantis_libvirt/" > /dev/null 2>&1 Apr 18 02:10:44 Atlantis CA Backup/Restore: Changing permissions on backup Apr 18 02:10:44 Atlantis CA Backup/Restore: Backing Up appData from /mnt/user/appdata/ to /mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected] Apr 18 02:10:45 Atlantis CA Backup/Restore: Using command: cd '/mnt/user/appdata/' && /usr/bin/tar -cvaf '/mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected]/CA_backup.tar' --exclude "alteria-plex/Library/Application Support/Plex Media Server/Metadata" --exclude "0-pihole" --exclude "alteria-postgres" --exclude "alteria-hydrus-server" * >> /var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log 2>&1 & echo $! > /tmp/ca.backup2/tempFiles/backupInProgress Apr 18 02:10:45 Atlantis CA Backup/Restore: Backup Complete Apr 18 02:10:45 Atlantis CA Backup/Restore: Verifying backup Apr 18 02:10:45 Atlantis CA Backup/Restore: Using command: cd '/mnt/user/appdata/' && /usr/bin/tar --diff -C '/mnt/user/appdata/' -af '/mnt/user/Backups/server_backups/atlantis_docker_appdata/[email protected]/CA_backup.tar' > /var/lib/docker/unraid/ca.backup2.datastore/appdata_backup.log & echo $! > /tmp/ca.backup2/tempFiles/verifyInProgress Apr 18 02:10:45 Atlantis kernel: br-55ef392db062: port 1(veth2d7642e) entered blocking state Apr 18 02:11:46 Atlantis kernel: br-55ef392db062: port 70(veth7449b16) entered forwarding state Apr 18 02:11:46 Atlantis CA Backup/Restore: ####################### Apr 18 02:11:46 Atlantis CA Backup/Restore: appData Backup complete Apr 18 02:11:46 Atlantis CA Backup/Restore: ####################### Apr 18 02:11:46 Atlantis CA Backup/Restore: Backup / Restore Completed Looking at the folder the backup should be in, it's empty. Any idea what happened?