Debaser

Members
  • Posts

    26
  • Joined

  • Last visited

Debaser's Achievements

Noob

Noob (1/14)

3

Reputation

  1. ooh i see, i think there was a new version released 2 days ago (0.4.4) where when i installed the mod pack via curse a few days ago it was 0.4.3. I see the container is using 0.4.4 so upgrading the client now to test thanks for the info edit: yep that was it, all good now, thanks!
  2. Hey, thanks for the container, makes it a lot easier than uploading mods to mineos. Running into an issue with AllTheMods7 container. It's up and running, I can see it, I can see my connection attempt in the logs, but I get an error `mismatched mod channel list` in the client when connecting. This seems to be a mismatch of mods or mod versions on the client vs the server. I don't play minecraft, just hosting a server for friends so I don't really know how to troubleshoot the mods. I downloaded the package and launched from the CurseForge app. Any easy way to compare what mods are mismatched? Thanks
  3. hey all, checking in here, system has been stable for about 2 weeks now. it seems the culprit may have been a pi-hole docker container i was running that was having issues to begin with on its own. ever since removing that container, it seems to have been running stable. going to keep monitoring for a while, but hopeful that may have been the solution. appreciate the help!
  4. So, had a few good days with zero issues, then looks like around 1am today (tuesday), looks like server crashed and went unresponsive again, had to power cycle through idrac interface Here's the syslog from the past 2 days. another kernel panic on the console output Any ideas? SyslogCatchAll-2020-06-29.txt SyslogCatchAll-2020-06-30.txt
  5. Thanks. I've deleted the docker image and re-installed my containers from community applications. I stepped away during the install of the containers. It looks like it finished, as all my containers are there, but it looks like sometime after the system had a kernel panic, and the system became unresponsive. I had to do a power cycle via the idrac. I've attached a screenshot of the unresponsive console with the kernel panic, as well as the syslog output from today. After power cycling, unraid is back up and everything seems to be running ok for now. I'll continue to monitor to see what happens SyslogCatchAll-2020-06-25.txt
  6. Got it, thanks. diagnostics attached. stevenet-diagnostics-20200625-1059.zip
  7. Hey guys, update here. I've been running the syslog server since last night. Everything as far as I can tell is still running smooth other than Grafana showing my Cache read/write at a constant 0. I have noticed this usually happens before I notice any major issues. Most of my docker containers are running in cache, so there should be some constant read/write activity I believe. I've attached my syslog output. There are plenty of critical errors such as the following 2020-06-24 21:22:55 Kernel.Info 192.168.1.204 Jun 24 21:22:53 stevenet kernel: BTRFS info (device loop2): no csum found for inode 4004 start 0 2020-06-24 21:22:55 Kernel.Critical 192.168.1.204 Jun 24 21:22:53 stevenet kernel: BTRFS critical (device loop2): corrupt leaf: root=7 block=12937510912 slot=43, bad key order, prev (18446744073709551606 128 12001943552) current (18446744073709551606 16 72057081092011947) 2020-06-24 21:22:55 Kernel.Critical 192.168.1.204 Jun 24 21:22:53 stevenet kernel: BTRFS critical (device loop2): corrupt leaf: root=7 block=12937510912 slot=43, bad key order, prev (18446744073709551606 128 12001943552) current (18446744073709551606 16 72057081092011947) 2020-06-24 21:22:55 Kernel.Info 192.168.1.204 Jun 24 21:22:53 stevenet kernel: BTRFS info (device loop2): no csum found for inode 4004 start 4096 After some googling, it appears loop2 would be my docker.img? Could this be the cause of my system crashing on me? What would my troubleshooting steps be here, delete my docker image and reinstall my containers? Or am I going down the wrong path here? SyslogCatchAll-2020-06-24.txt SyslogCatchAll-2020-06-25.txt
  8. Thanks guys for the responses. I've gone and set up a remote syslog server, as well as tailing the syslog in my idrac console. I'll post back with any information on the next crash. @jonp - this issue did just start happening in the past few weeks, after more than a year running perfectly stable. I did notice the cache ssd temperature warnings and realized that unraid has some generic temperate warning levels. I found my SSDs should warn around 60c and critical around 75, so I have adjusted those levels, so heat does not seem to be the issue. I haven't made any hardware changes recently, and according to grafana and the unraid dashboard, my resources don't seem to be strapped. i've been transcoding a large video library from x264 to x265 using tdarr since January which may be putting some strain on the CPUs. I've dropped it from 3 workers to 1 to lessen the strain, but it does not seem like anything has changed.
  9. so a few times since my last post about a week ago, my server has needed a hard reboot 2-3 times. it's currently in a state where some containers and services are working, but some are not. not sure what is going on. i'm pretty sure if i try a graceful shutdown at this point it will just hang and require another hard reboot pages in unraid such as docker, dashboard, etc are unresponsive or do not load. others such as Main, Shares seem to be loading fine anyone have any advice on how to troubleshoot this?
  10. gotcha, thanks for the info. so you're saying the SSD trim isn't likely what is causing these issues with unraid? it's been stable for the last 24 hours or so, but has been pretty unstable for the past week
  11. Thanks for the info. I'll have to take a look when I get home. I thought I read that on the r720XD the onboard sata is disabled how important is it to run trim on the SSDs? would I be better off disabling the trim plugin if I can't use the onboard SATA?
  12. sorry, diagnostics attached stevenet-diagnostics-20200614-1835.zip
  13. Hey all, been having some strange issues lately that I'm not even sure how to properly describe. Over the past 2 weeks or so, I've been experiencing issues like my array becoming unresponsive, webui not loading all elements, docker service becoming stuck, unable to reboot gracefully. I've experienced about 3-4 times in this period where I've had to power cycle the server via idrac, no graceful shutdown. I can use the idrac console or ssh into unraid to run a reboot command, but it just says the server is going down for a reboot, and just sits there until I power cycle. I believe the issue is with my cache drives. I have a 2x800gb ssd cache pool. I have been seeing that they have been running hot often, but usually coming back down in temp shortly after I get the temperature notifications. In my /root directiry, I see a file called dead.letter. Here are the contents of that file Event: Unraid Parity check Subject: Notice [STEVENET] - Parity check started Description: Size: 8 TB Importance: warning fstrim: /mnt/cache: FITRIM ioctl failed: Remote I/O error fstrim: /mnt/cache: FITRIM ioctl failed: Remote I/O error fstrim: /mnt/cache: FITRIM ioctl failed: Remote I/O error fstrim: /mnt/cache: FITRIM ioctl failed: Remote I/O error When I console in through idrac, I see some messages like this (attached) My SMART checks have all passed when I last ran them, which was a day or two ago, after I started experiencing these issues. Is this a sign of my cache drive(s) failing, or possibly something else? Not sure which steps to take next. Thanks
  14. hmm ok so i'm stupid, the drive just isn't mounted apparently looking at `df -h` i don't see it listed anywhere, but when i run a `fdisk -l` i see it i was thinking the disk and partitons would be automatically be mounted. should i be mounting this manually with `/etc/fstab`?
  15. Hey there, trying to figure out how to pre-allocate the full disk size on my Ubuntu 18.04 server VM. I'm running unifi nvr and there's a setting to keep at least X space free, and the lowest setting is 10gb. my VM only has ~500mb free because it has an expanding disk size it seems. Trying to figure the best way to go about this. I suppose I could mount one of my unraid shares, but i haven't been able to get that to work either. any ideas?