Jump to content

Macj72x

Members
  • Posts

    29
  • Joined

Posts posted by Macj72x

  1. My server has been having intermittent issues where the gui and ssh will become no longer accessible until I reboot the server.  I received a notification that it froze at around 11ish today 4/17.  I set up a remote sys log server and attached the results.  Could someone provide me some guidance on what may be happening or how I can troubleshoot this?  I don't see anything in the logs that stick out to me.

    syslog-192.168.3.7.log

  2. I am having trouble getting the array to stop.  I keep getting these errors continuously:

     

    Feb 7 15:52:11 ChunkyMonkey emhttpd: Unmounting disks...

    Feb 7 15:52:11 ChunkyMonkey emhttpd: shcmd (310): /usr/sbin/zpool export cache-system

    Feb 7 15:52:11 ChunkyMonkey root: cannot unmount '/mnt/cache-system': pool or dataset is busy

    Feb 7 15:52:11 ChunkyMonkey emhttpd: shcmd (310): exit status: 1

    Feb 7 15:52:11 ChunkyMonkey emhttpd: Retry unmounting disk share(s)...

    Feb 7 15:52:16 ChunkyMonkey emhttpd: Unmounting disks...

     

    I followed this link:

    and ran losetup to receive these results:


    NAME       SIZELIMIT OFFSET AUTOCLEAR RO BACK-FILE                                  DIO LOG-SEC
    /dev/loop1         0      0         1  1 /boot/bzfirmware                             0     512
    /dev/loop2         0      0         1  0 /mnt/cache-system/system/docker/docker.img   0     512
    /dev/loop0         0      0         1  1 /boot/bzmodules                              0     512

     

    I ran umount /dev/loop2 to force the array to stop.

     

    Are there any diagnosing steps to prevent this from happening again?

     

    I've attached my diagnostics.

    chunkymonkey-diagnostics-20240207-1556.zip

  3. So I solved it.  I was thinking you change the setting at the RAID controller configuration to go to HBA mode, but that didn't work and caused me more issues.  After lots of tinkering and get my drives to show again, I found the best solution.  I had to go into the configuration of each individual drive and change it to non-raid mode then click go.

  4. When I try to start, stop, restart, pause, force stop, hibernate, remove VM, or remove VM and Disks a container or VM it doesn't do anything and will just sit there with the menu highlighted.

    image.png.2ed7d737f7d540e0f951af3fe3e60b9c.png

     

    Without clicking everything, most all the other menu items seem to work find including start with console(VNC) for the vms. I've tried using different browsers.  This seems to be a relatively new development maybe in one of the last few updates.  I have no clue where to start with diagnosing this.  Please see my attached diagnostics.

     

    chunkymonkey-diagnostics-20231207-1449.zip

    • Upvote 1
  5. I redid the flash drive.  I didn't get a boot error the two times I tried booting it so fingers crossed that is fixed, but my appdata share ownership keeps changing to unknown.  Any advice on how to diagnose that?

  6. I lately have been having this kernel panic pop up when I need to restart unraid:

    unraid.thumb.jpg.721453ab50a1e1f93d70f8446d447597.jpg

    Sometimes I can restart and it will go past it and sometimes it takes several restarts.  Can anyone decode this for a layman?

     

    Once I get Unraid to boot, sometimes I will be missing some of my cache nvme drives.  This probably happens 1 in 10 boots and I will need to restart for them to appear again.

     

    Then once I get it going with all the drives, I will have all my appdata owership change from nobody to unknown.  This causes all my docker containers to not work until I run a script a few times switching it back.  The appdata share is on a different cache pool than the nvme drives that disappear.

     

    I feel like this may all be related and a hardware issue, but I'm not sure where to start to diagnose this.  Any help?

    chunkymonkey-diagnostics-20230809-1143.zip

  7. As the title says, I keep having ownership changes with my entire appdata share and only that share.  It causes issues with all my containers.  I happen to catch it in the middle of a change and copied the attached diag file.  It started changing sometime around 11-12am today(6/21).  I don't see anything that stands out in my log file.  Can someone more experienced look through and see if anything stands out that may be causing this?

    chunkymonkey-diagnostics-20230621-1158.zip

  8. As the title says, I keep having ownership changes with my entire appdata share and only that share.  Obviously it causes issues with all my containers when this happens and I have to run a Fix Appdata Permissions script to fix it. I currently have to run that several times a week.  The last time it happened (some time right before 2023-04-05 14:57:46) this is the only message I saw before it changed:

    Apr  5 14:56:53 ChunkyMonkey kernel: docker0: port 13(vethf0ca75c) entered disabled state
    Apr  5 14:56:53 ChunkyMonkey kernel: veth3e9fb99: renamed from eth0
    Apr  5 14:56:53 ChunkyMonkey  avahi-daemon[14646]: Interface vethf0ca75c.IPv6 no longer relevant for mDNS.
    Apr  5 14:56:53 ChunkyMonkey  avahi-daemon[14646]: Leaving mDNS multicast group on interface vethf0ca75c.IPv6 with address fe80::54a2:2aff:fe29:9011.
    Apr  5 14:56:53 ChunkyMonkey kernel: docker0: port 13(vethf0ca75c) entered disabled state
    Apr  5 14:56:53 ChunkyMonkey kernel: device vethf0ca75c left promiscuous mode
    Apr  5 14:56:53 ChunkyMonkey kernel: docker0: port 13(vethf0ca75c) entered disabled state
    Apr  5 14:56:53 ChunkyMonkey  avahi-daemon[14646]: Withdrawing address record for fe80::54a2:2aff:fe29:9011 on vethf0ca75c.

    This doesn't seems like something that would cause that to my untrained eye, but is all I see.  I've also noticed it seems to happen around the time that appdata backup runs.  This is exhausting as I'm fairly new to Unraid and want to get this thing to be stable, but I'm about to give up.  Please help me get this thing running properly.

    chunkymonkey-diagnostics-20230406-0829.zip

  9. Hopefully I don't need to make a new tread, but I'm still having trouble with mover.  Since I made this post, I have built a separate cache pool for my default shares and those are set to prefer.  All my media and basically everything else is goes into a separate media cache pool and is set to Yes:cache-media. That media pool keeps getting completely full and I have to shut everything down and wait a few hours for it to start emptying.

     

    Can someone tell me what is going on here or where I can search the logs to find out why stuff isn't moving? 

    chunkymonkey-diagnostics-20230319-1922.zip

  10. 35 minutes ago, trurl said:

    Now, I have downloads that need post-processing, and I have enough hardware that I can download to a cached share, post-process to my "fast" pool, where it gets moved ('arr) to the final destination cached share.

    Could you break this down a little more? You use two cache pools for your media downloads or just one of the two? What is the path that your media downloads take because that obviously is my issue. Sounds like a great recommendation to break up my cache pool into two.

  11. 56 minutes ago, trurl said:

     

    What do you get from command line with this?

    du -h -d 1 /mnt/cache

     

    27G     /mnt/cache/appdata
    49G     /mnt/cache/domains
    71G     /mnt/cache/system
    0       /mnt/cache/temp
    3.3T    /mnt/cache/data
    50G     /mnt/cache/backups
    3.5T    /mnt/cache

     

    Mover seems to be moving now that both docker and vm is off, but it wasn't earlier in the day when just docker was off.  Like I said earlier, the only vm I'm running right now is Homeassistant which shouldn't be touching my media share.

  12.  

     

    I have changed the default shares to only.  I have deleted the other two shares listed as prefer as they didn't have any important data on them anyway.

     

    I ran the mover and nothing has moved off the cache since you last posted approximately 2 hours ago.  Other than the default shares, all shares are either listed as Yes:Cache or No with nothing on the cache.  Computing all the share size showed that a vast majority of the cache data is from my media download/storage drive (named data).

  13. While I still would like to know if I’m using the cache pool in a way that isn’t recommended, I still think something else is wrong with my mover or the way it is configured.  I’ve turned off docker. I now only have my Homeassistant VM running.  I’ve checked my open files and it doesn’t appear that I have anything with my media files open. 65847BFD-F021-4C00-9FF1-87C5529449CE.thumb.jpeg.71616d2137331649f8a8709064a06dad.jpeg
     

    After this, I’ve invoked mover and it has been running for over and hour and has hardly moved anything. I started with about 250gb space remaining and am now at 267gb. Any guesses where to start?BF4F52A7-23AE-412E-9207-F9AC2A7315E2.thumb.jpeg.c129851539240abe55ad5e2b9dcef5da.jpeg

×
×
  • Create New...