Jump to content
  • [6.12.6] Server hangs and hard rebot is the only usable option


    sn0wd0g
    • Urgent

    The server freezes every few days and can only be restarted from the button.

    The same behavior is happening since 6.12.1 with all versions until 6.12.6.

    I have tested all releases from 6.12.1 until the latest and have always reverted back to 6.11.5 which seems very stable.

     

     

    I cannot generate the diagnostics as the server is not responsive.

    This is the current status according to "top"
     

    top - 18:07:33 up 2 days, 23:16,  4 users,  load average: 536.77, 541.29, 476.42
    Tasks: 1183 total,  17 running, 1166 sleeping,   0 stopped,   0 zombie
    %Cpu(s):  1.1 us, 93.4 sy,  0.0 ni,  0.0 id,  5.2 wa,  0.0 hi,  0.3 si,  0.0 st
    MiB Mem :  31832.8 total,    920.7 free,  12348.8 used,  18563.3 buff/cache
    MiB Swap:      0.0 total,      0.0 free,      0.0 used.  17504.5 avail Mem
    
      PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+ COMMAND
      163 root      20   0       0      0      0 R  71.1   0.0 611:48.68 kswapd0
     4640 root      20   0       0      0      0 R  54.3   0.0   9:28.53 kworker/u24:8+loop2
    31161 root      20   0       0      0      0 R  53.3   0.0   8:40.87 kworker/u24:7+loop2
    20072 root      20   0       0      0      0 R  48.7   0.0   9:13.94 kworker/u24:4+loop2
    26796 root      20   0       0      0      0 R  47.4   0.0   7:45.40 kworker/u24:3+loop2
    30526 root      20   0       0      0      0 R  46.4   0.0   8:01.16 kworker/u24:10+loop2
    18845 root      20   0       0      0      0 I  44.4   0.0   7:31.15 kworker/u24:5-btrfs-endio
    11327 root      20   0       0      0      0 R  40.5   0.0   8:01.96 kworker/u24:21+loop2
    23368 root      20   0       0      0      0 R  40.1   0.0   7:38.71 kworker/u24:26+loop2
     9418 root      20   0       0      0      0 R  39.8   0.0   4:52.91 kworker/u24:18+loop2
     9318 root      20   0       0      0      0 I  39.5   0.0   9:07.26 kworker/u24:13-btrfs-endio
    23350 root      20   0       0      0      0 R  38.8   0.0   7:40.69 kworker/u24:25+loop2
    11799 root      20   0       0      0      0 R  35.9   0.0   8:20.71 kworker/u24:6+loop2
    11324 root      20   0       0      0      0 I  35.2   0.0   8:11.04 kworker/u24:20-btrfs-endio
    23839 root      20   0       0      0      0 I  34.9   0.0   4:46.69 kworker/u24:28-btrfs-endio
    28572 root      20   0       0      0      0 I  34.9   0.0   8:26.47 kworker/u24:1-btrfs-endio
    10287 root      20   0       0      0      0 I  34.2   0.0   9:01.77 kworker/u24:0-btrfs-endio
    16594 root      20   0       0      0      0 R  32.2   0.0   8:56.59 kworker/u24:14+btrfs-endio
    23779 root      20   0       0      0      0 I  30.9   0.0   5:43.96 kworker/u24:27-btrfs-endio
    21281 root      20   0       0      0      0 I  30.6   0.0   6:03.77 kworker/u24:12-btrfs-endio
    19565 root      20   0       0      0      0 I  29.9   0.0   8:21.53 kworker/u24:24-btrfs-endio
     9319 root      20   0       0      0      0 I  28.9   0.0   9:18.39 kworker/u24:17-btrfs-endio
    18390 root      20   0       0      0      0 R  26.0   0.0  10:05.56 kworker/u24:9+loop2
    11333 root      20   0       0      0      0 R  21.1   0.0   8:33.80 kworker/u24:23+loop2
      553 root      20   0       0      0      0 I  14.8   0.0   8:30.42 kworker/u24:2-btrfs-endio-meta

    Syslog is also attached

    syslog

     

    Memtest was successful.

    Docker is running with ipvlan.

     

    Diag generated after system reboot.

     

     

     

    zabox-diagnostics-20240104-1910 2.zip




    User Feedback

    Recommended Comments

    I had a new crash today, as Ive switched back from 6.11.5 to 6.12.6. I've got a new diagnostics just when the load was rising.

    I was checking the docker stats right at that moment, and all containers seemed ok from the stats overview. none of them was created a cpu load and memory was within the limits.

    zabox-diagnostics-20240108-1006.zip

    Edited by sn0wd0g
    Link to comment

    I too have had been fighting this issue for a while now and after alot of time and testing... I have narrowed the issue down to my btrfs DiskPool... I ended up moving everything off of it due to this problem... not happy

    Link to comment


    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Restore formatting

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Status Definitions

     

    Open = Under consideration.

     

    Solved = The issue has been resolved.

     

    Solved version = The issue has been resolved in the indicated release version.

     

    Closed = Feedback or opinion better posted on our forum for discussion. Also for reports we cannot reproduce or need more information. In this case just add a comment and we will review it again.

     

    Retest = Please retest in latest release.


    Priority Definitions

     

    Minor = Something not working correctly.

     

    Urgent = Server crash, data loss, or other showstopper.

     

    Annoyance = Doesn't affect functionality but should be fixed.

     

    Other = Announcement or other non-issue.

×
×
  • Create New...