Jump to content
  • [6.8.3] shfs error results in lost /mnt/user


    JorgeB
    • Minor

    There are several reports in the forums of this shfs error causing /mnt/user to go away:

     

    May 14 14:06:42 Tower shfs: shfs: ../lib/fuse.c:1451: unlink_node: Assertion `node->nlookup > 1' failed.

     

    Rebooting will fix it, until it happens again, I remember seeing at least 5 or 6 different users with the same issue in the last couple of months, it was reported here that it's possibly this issue:

     

    https://github.com/libfuse/libfuse/issues/128

     

    Attached diags from latest occurrence.

     

     

     

    tower-diagnostics-20200514-1444.zip

    • Upvote 4



    User Feedback

    Recommended Comments



    dlandon

    Posted

    7 hours ago, robertklep said:

    For me, the best way to trigger the issue was using NFS, but even with NFS completely turned off and using SMB it happened.

    Please make this happen and post a diagnostics.

    robertklep

    Posted (edited)

    2 hours ago, dlandon said:

    Please make this happen and post a diagnostics.

     

    Posted those about a month ago, then found this thread about what I believe to be the same issue (or at least related) that started back in 2018.

     

    As for the Docker documentation, that's just general information, at least I don't see anything specific to what you shouldn't do to prevent `shfs` from getting into trouble.

    Edited by robertklep
    dlandon

    Posted

    1 hour ago, robertklep said:

    Posted those about a month ago, then found this thread about what I believe to be the same issue (or at least related) that started back in 2018.

    Can you link me to your diagnostics?

     

    1 hour ago, robertklep said:

    As for the Docker documentation, that's just general information, at least I don't see anything specific to what you shouldn't do to prevent `shfs` from getting into trouble.

    In general, misconfiguring docker access on mount points can cause many different problems.

    robertklep

    Posted

    53 minutes ago, dlandon said:

    Can you link me to your diagnostics?

     

    They're attached to the post: https://forums.unraid.net/applications/core/interface/file/attachment.php?id=284784&key=276c1e0db7ed3ce86c8547d0c185730b

     

    53 minutes ago, dlandon said:

    In general, misconfiguring docker access on mount points can cause many different problems.

     

    Fair enough, but in all my time using Docker on various systems I've never come across a filesystem driver hanging itself up over it.

    • Like 1
    grenskul

    Posted

    Can we get this reclassified to urgent?

    This is a problem that forces you to restart your server. If it happens and you don't notice you WILL lose data.

    • Upvote 1
    dlandon

    Posted

    1 hour ago, robertklep said:

    Here is what I'm seeing in your log:

    Dec 13 11:19:15 Unraid network: update services: 30s
    Dec 13 11:19:16 Unraid root: Installing /boot/extra packages
    Dec 13 11:19:18 Unraid root: Installing: vim-8.2.4256-x86_64-1: Vi IMproved ......................................................................... [  37M]
    Dec 13 11:19:18 Unraid root: Installing: libsodium-1.0.18-x86_64-3: Sodium crypto library ........................................................... [ 620K]
    Dec 13 11:19:18 Unraid root: Installing: mosh-1.4.0-x86_64-4cf: MObile SHell server and client ...................................................... [ 860K]
    Dec 13 11:19:19 Unraid root: Installing: protobuf-21.12-x86_64-1cf: Google's data interchange format ................................................ [  17M]

    and

    Dec 13 14:11:18 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.5:778 for /mnt/user/gaming (/mnt/user/gaming)
    Dec 13 14:11:23 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.5:777 for /mnt/user/gaming (/mnt/user/gaming)
    Dec 13 14:44:15 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:773 for /mnt/user/movies (/mnt/user/movies)
    Dec 13 14:44:15 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:779 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 16:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:790 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 16:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:794 for /mnt/user/movies (/mnt/user/movies)
    Dec 13 16:30:24 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:812 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 20:28:37 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:818 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 21:50:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:824 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 21:50:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:828 for /mnt/user/movies (/mnt/user/movies)
    Dec 13 21:52:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:834 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 21:52:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:838 for /mnt/user/movies (/mnt/user/movies)
    Dec 13 22:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:857 for /mnt/user/tv (/mnt/user/tv)
    Dec 13 22:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:861 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 00:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:881 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 00:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:885 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 00:01:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:899 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 00:01:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:903 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 02:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:921 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 02:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:925 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 02:01:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:931 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 02:01:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:935 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 04:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:953 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 04:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:957 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 06:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:975 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 06:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:979 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 08:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:1002 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 08:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:1006 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 08:20:10 Unraid monitor: Stop running nchan processes
    Dec 14 08:44:29 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:702 for /mnt/user/movies (/mnt/user/movies)
    Dec 14 08:44:29 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:707 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 10:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:717 for /mnt/user/tv (/mnt/user/tv)
    Dec 14 10:00:00 Unraid rpc.mountd[5019]: authenticated mount request from 192.168.23.15:721 for /mnt/user/movies (/mnt/user/movies)

    and then the system was told to shut down

    Dec 15 09:46:09 Unraid shutdown[11136]: shutting down for system halt
    Dec 15 09:46:09 Unraid init: Switching to runlevel: 0
    Dec 15 09:46:09 Unraid init: Trying to re-exec init
    Dec 15 09:47:41 Unraid root: Status of all loop devices
    Dec 15 09:47:41 Unraid root: /dev/loop1: [2049]:12 (/boot/bzfirmware)
    Dec 15 09:47:41 Unraid root: /dev/loop2: [2305]:6442451073 (/mnt/disk1/system/libvirt/libvirt.img)
    Dec 15 09:47:41 Unraid root: /dev/loop0: [2049]:10 (/boot/bzmodules)
    Dec 15 09:47:41 Unraid root: Active pids left on /mnt/*
    Dec 15 09:47:42 Unraid root: Cannot stat /mnt/user: Software caused connection abort
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/740/fd/8: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/740/fd/9: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/740/fd/10: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/740/fd/14: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/740/fd/16: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/858/fd/3: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/858/fd/7: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/7019/fd/8: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/7023/fd/8: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/7024/fd/8: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid root: Cannot stat file /proc/7025/fd/8: Transport endpoint is not connected
    Dec 15 09:47:42 Unraid kernel: ------------[ cut here ]------------
    Dec 15 09:47:42 Unraid kernel: nfsd: non-standard errno: -103
    Dec 15 09:47:42 Unraid kernel: WARNING: CPU: 2 PID: 5015 at fs/nfsd/nfsproc.c:909 nfserrno+0x45/0x51 [nfsd]
    Dec 15 09:47:42 Unraid kernel: Modules linked in: tcp_diag inet_diag bluetooth ecdh_generic ecc tls xt_CHECKSUM ipt_REJECT nf_reject_ipv4 ip6table_mangle ip6table_nat iptable_mangle vhost_net vhost vhost_iotlb xt_comment xt_connmark xt_mark nft_compat nf_tables wireguard curve25519_x86_64 libcurve25519_generic libchacha20poly1305 chacha_x86_64 poly1305_x86_64 ip6_udp_tunnel udp_tunnel libchacha tun veth xt_nat xt_tcpudp xt_conntrack xt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 xt_addrtype br_netfilter xfs nfsd auth_rpcgss oid_registry lockd grace sunrpc md_mod zfs(PO) zunicode(PO) zzstd(O) zlua(O) zavl(PO) icp(PO) zcommon(PO) znvpair(PO) spl(O) ip6table_filter ip6_tables iptable_filter ip_tables x_tables efivarfs macvtap macvlan tap af_packet 8021q garp mrp bridge stp llc igb intel_rapl_msr intel_rapl_common iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm crct10dif_pclmul crc32_pclmul crc32c_intel
    Dec 15 09:47:42 Unraid kernel: ghash_clmulni_intel sha512_ssse3 sha256_ssse3 ipmi_ssif sha1_ssse3 wmi_bmof ast drm_vram_helper drm_ttm_helper ttm aesni_intel crypto_simd cryptd drm_kms_helper rapl intel_cstate drm intel_uncore i2c_i801 i2c_algo_bit mei_me agpgart ahci syscopyarea sysfillrect sysimgblt i2c_smbus fb_sys_fops i2c_core libahci mei nvme cp210x pl2303 input_leds joydev led_class usbserial acpi_ipmi intel_pch_thermal nvme_core thermal fan video wmi ipmi_si backlight intel_pmc_core acpi_tad button unix [last unloaded: igb]
    Dec 15 09:47:42 Unraid kernel: CPU: 2 PID: 5015 Comm: nfsd Tainted: P           O       6.1.64-Unraid #1
    Dec 15 09:47:42 Unraid kernel: Hardware name: Supermicro Super Server/X11SCL-IF, BIOS 2.2 10/27/2023
    Dec 15 09:47:42 Unraid kernel: RIP: 0010:nfserrno+0x45/0x51 [nfsd]
    Dec 15 09:47:42 Unraid kernel: Code: c3 cc cc cc cc 48 ff c0 48 83 f8 26 75 e0 80 3d dd c9 05 00 00 75 15 48 c7 c7 b5 c2 d9 a0 c6 05 cd c9 05 00 01 e8 01 39 30 e0 <0f> 0b b8 00 00 00 05 c3 cc cc cc cc 48 83 ec 18 31 c9 ba ff 07 00
    Dec 15 09:47:42 Unraid kernel: RSP: 0000:ffffc9000155fde8 EFLAGS: 00010286
    Dec 15 09:47:42 Unraid kernel: RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000027
    Dec 15 09:47:42 Unraid kernel: RDX: 0000000000000002 RSI: ffffffff820d7e01 RDI: 00000000ffffffff
    Dec 15 09:47:42 Unraid kernel: RBP: ffff88814e140180 R08: 0000000000000000 R09: ffffffff82245f10
    Dec 15 09:47:42 Unraid kernel: R10: 00007fffffffffff R11: ffffffff82969256 R12: 0000000000000001
    Dec 15 09:47:42 Unraid kernel: R13: 0000000000000000 R14: ffff88814f6dc0c0 R15: ffffffffa0dbf6c0
    Dec 15 09:47:42 Unraid kernel: FS:  0000000000000000(0000) GS:ffff88845ed00000(0000) knlGS:0000000000000000
    Dec 15 09:47:42 Unraid kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    Dec 15 09:47:42 Unraid kernel: CR2: 000014a2aa5dbbd3 CR3: 0000000241028002 CR4: 00000000003706e0
    Dec 15 09:47:42 Unraid kernel: Call Trace:
    Dec 15 09:47:42 Unraid kernel: <TASK>
    Dec 15 09:47:42 Unraid kernel: ? __warn+0xab/0x122
    Dec 15 09:47:42 Unraid kernel: ? report_bug+0x109/0x17e
    Dec 15 09:47:42 Unraid kernel: ? nfserrno+0x45/0x51 [nfsd]
    Dec 15 09:47:42 Unraid kernel: ? handle_bug+0x41/0x6f
    Dec 15 09:47:42 Unraid kernel: ? exc_invalid_op+0x13/0x60
    Dec 15 09:47:42 Unraid kernel: ? asm_exc_invalid_op+0x16/0x20
    Dec 15 09:47:42 Unraid kernel: ? nfserrno+0x45/0x51 [nfsd]
    ### [PREVIOUS LINE REPEATED 1 TIMES] ###
    Dec 15 09:47:42 Unraid kernel: nfsd_access+0xac/0xf1 [nfsd]
    Dec 15 09:47:42 Unraid kernel: nfsd3_proc_access+0x78/0x88 [nfsd]
    Dec 15 09:47:42 Unraid kernel: nfsd_dispatch+0x1a6/0x262 [nfsd]
    Dec 15 09:47:42 Unraid kernel: svc_process_common+0x32f/0x4df [sunrpc]
    Dec 15 09:47:42 Unraid kernel: ? ktime_get+0x35/0x49
    Dec 15 09:47:42 Unraid kernel: ? nfsd_svc+0x2b6/0x2b6 [nfsd]
    Dec 15 09:47:42 Unraid kernel: ? nfsd_shutdown_threads+0x5b/0x5b [nfsd]
    Dec 15 09:47:42 Unraid kernel: svc_process+0xc7/0xe4 [sunrpc]
    Dec 15 09:47:42 Unraid kernel: nfsd+0xd5/0x155 [nfsd]
    Dec 15 09:47:42 Unraid kernel: kthread+0xe4/0xef
    Dec 15 09:47:42 Unraid kernel: ? kthread_complete_and_exit+0x1b/0x1b
    Dec 15 09:47:42 Unraid kernel: ret_from_fork+0x1f/0x30
    Dec 15 09:47:42 Unraid kernel: </TASK>
    Dec 15 09:47:42 Unraid kernel: ---[ end trace 0000000000000000 ]---

    it looks like without unmounting any of the mounts or deling with Unriad going offline.

     

    and

    Dec 15 09:47:42 Unraid kernel: traps: mariadbd[17135] general protection fault ip:148623228898 sp:1485faffe850 error:0 in libc.so.6[148623228000+195000]

     

    So this is what I'd do in the order I recommend:

    • Remove the extra packages being loaded.
    • Remove two unknown plugins:
      • un-get.plg - 2023.11.12  (Unknown to Community Applications).
      • unraid-tmux.plg - plugin: version attribute not present  (Unknown to Community Applications).
    • Find out what is causing the mariadb issue.  Granted it occurred after the shutdown was initiated, but why is it not being mnaged by the shutdown?  e.g. a Docker Container that would shut down properly.
    • Check your mount parameters for the NFS mounts.  UD uses these mount parameters:
      • soft,relatime,retrans=4,timeo=300.
    • Manage the mounts on a shutdown.  It looks like the NFS mounts are remote mounts from other Linux boxes.  Leaving them mounted when Unraid is shutting down is not a good idea.  Have the clients manage them a bit better.  The above NFS mount options may help.
    • Cut down on the number of NFS mounts.  You are asking Linux to manage an awful lot over a 1GB network.  It may be choking.
    • Go back to a more basic system and start building it back a little at a time and check performance as you go.
    dlandon

    Posted

    1 hour ago, grenskul said:

    Can we get this reclassified to urgent?

    This is a problem that forces you to restart your server. If it happens and you don't notice you WILL lose data.

    I'm here working on the issue as best I can.  I need more information like disgnostics so I can help with troubleshooting.  Unfortunately, we don't see anything common that can be acted on.  If we can find a common theme, we can turn it into an action item for the LT team.

    robertklep

    Posted (edited)

    On 1/13/2024 at 8:41 PM, dlandon said:

    Here is what I'm seeing in your log:

    ...

    and

    ...

    and then the system was told to shut down

    ...

    it looks like without unmounting any of the mounts or deling with Unriad going offline.

     

    So to clarify what actually happened: `shfs` deadlocked for some reason. I was running htop (see screenshot in my post) which showed some random postfix process hanging in a D state.

     

    At Dec 15, 09:46:09, I then initiated a shutdown, which got stuck because `/mnt/user` was borked. I kill-9'd ("voluntary" signals didn't work, another clue that the process was stuck in some unforeseen state) the shfs processes that I could see in the htop output causing a ton of "Transport endpoint is not connected" errors and very likely also the mariadb error that you mention (I only just noticed that the nfsd error that I initially thought could be cause of the hang actually came after I killed shfs, so that's just caused by `/mnt/user` disappearing).

     

    Killing shfs caused disk I/O to get unstuck enough for Unraid to actually write diagnostics and finish the shutdown.

     

    Note that up until the point where I started the shutdown, nothing was being logged. Between 08:00AM (when Kodi NFS-mounted a share to do a library scan) and 09:46:09, something happened that caused shfs to deadlock.

     

    On 1/13/2024 at 8:41 PM, dlandon said:

    So this is what I'd do in the order I recommend:

    • Remove the extra packages being loaded.
    • Remove two unknown plugins:
      • un-get.plg - 2023.11.12  (Unknown to Community Applications).
      • unraid-tmux.plg - plugin: version attribute not present  (Unknown to Community Applications).
    • Find out what is causing the mariadb issue.  Granted it occurred after the shutdown was initiated, but why is it not being mnaged by the shutdown?  e.g. a Docker Container that would shut down properly.
    • Check your mount parameters for the NFS mounts.  UD uses these mount parameters:
      • soft,relatime,retrans=4,timeo=300.
    • Manage the mounts on a shutdown.  It looks like the NFS mounts are remote mounts from other Linux boxes.  Leaving them mounted when Unraid is shutting down is not a good idea.  Have the clients manage them a bit better.  The above NFS mount options may help.
    • Cut down on the number of NFS mounts.  You are asking Linux to manage an awful lot over a 1GB network.  It may be choking.
    • Go back to a more basic system and start building it back a little at a time and check performance as you go.

     

    To debug this issue over the last few months, I've tried everything already. No additional packages, disabling NFS entirely, etc. Nothing fixed it. I've also had this problem from the moment I started using Unraid, so with a very fresh and basic system.

     

    Also, there aren't a lot of NFS mounts (probably about 5 at most). You see a lot of mount requests, but that's because Kodi manages NFS sources on demand, closing them when its done with them (which you don't see being logged).

     

    Since I got fed up with this instability I moved from Unraid to Ubuntu which, with the exact same workload in terms of NFS, Docker, etc, now has an uptime of 3 weeks, which with Unraid would have been "an exceptional run". I'm perfectly willing to accept that something in the way I use my server doesn't fit with Unraid, but on the other hand, I'm not the only person that's running into these issues, judging by this thread. I'm not willing to dumb down my server to fit the OS, so instead I changed OS.

    Edited by robertklep
    Subasically

    Posted

    I too am experiencing this issue using VSCode to SSH into my server to modify some bash file. A restart does fix it but I do not want to restart every time I want to edit some file using VSCode from a different computer.

     

    I am on 6.12.6.

    basic-cable-diagnostics-20240131-0848.zip

    dlandon

    Posted

    1 hour ago, Subasically said:

    I too am experiencing this issue using VSCode to SSH into my server to modify some bash file. A restart does fix it but I do not want to restart every time I want to edit some file using VSCode from a different computer.

    I see several things, but cannot offer any explanaton.  In the log:

    Jan 31 08:34:23 BASIC-CABLE shfs: error: strcpy_share_path, 455: No such file or directory (2): path too long: /CONTAINER ID   IMAGE                                 COMMAND                  CREATED         STATUS                    PORTS                                                                                                                       NAMES                    SIZE

     

    and the output of df shows:

    Filesystem            Size  Used Avail Use% Mounted on
    rootfs                 16G  1.7G   15G  11% /
    tmpfs                  32M  1.5M   31M   5% /run
    /dev/sda1              29G  981M   28G   4% /boot
    overlay                16G  1.7G   15G  11% /lib
    overlay                16G  1.7G   15G  11% /usr
    devtmpfs              8.0M     0  8.0M   0% /dev
    tmpfs                  16G     0   16G   0% /dev/shm
    tmpfs                 128M  396K  128M   1% /var/log
    tmpfs                 1.0M     0  1.0M   0% /mnt/disks
    tmpfs                 1.0M     0  1.0M   0% /mnt/remotes
    tmpfs                 1.0M     0  1.0M   0% /mnt/addons
    tmpfs                 1.0M     0  1.0M   0% /mnt/rootshare
    /dev/md1p1            5.5T  5.3T  168G  98% /mnt/disk1
    /dev/md2p1            5.5T  5.3T  175G  97% /mnt/disk2
    /dev/md3p1            5.5T  5.3T  171G  97% /mnt/disk3
    /dev/md4p1            5.5T  5.3T  177G  97% /mnt/disk4
    /dev/md5p1            3.7T  3.4T  341G  91% /mnt/disk5
    /dev/md6p1            3.7T  3.4T  322G  92% /mnt/disk6
    /dev/md7p1            3.7T  3.4T  338G  91% /mnt/disk7
    /dev/md8p1            3.7T  3.4T  325G  92% /mnt/disk8
    cache                 1.7T  256K  1.7T   1% /mnt/cache
    /dev/sdb1             257G  249G  7.6G  98% /mnt/docker-cache
    /dev/sdc1             489G   36G  454G   8% /mnt/plex-cache
    cache/data            1.7T  512K  1.7T   1% /mnt/cache/data
    cache/isos            1.7T   32G  1.7T   2% /mnt/cache/isos
    cache/unraid_scripts  1.7T  203M  1.7T   1% /mnt/cache/unraid_scripts
    cache/system          1.7T   30G  1.7T   2% /mnt/cache/system
    cache/domains         1.7T  128K  1.7T   1% /mnt/cache/domains
    cache/Music           1.7T   29G  1.7T   2% /mnt/cache/Music
    cache/Plex DVR        1.7T  1.8G  1.7T   1% /mnt/cache/Plex DVR
    cache/vm-disks        1.7T   16G  1.7T   1% /mnt/cache/vm-disks
    cache/temp            1.7T  128K  1.7T   1% /mnt/cache/temp
    cache/appdata         1.7T  5.1G  1.7T   1% /mnt/cache/appdata
    cache/iCloud-Photos   1.7T  7.7G  1.7T   1% /mnt/cache/iCloud-Photos
    cache/Pre-Rolls       1.7T  451M  1.7T   1% /mnt/cache/Pre-Rolls
    shfs                   37T   35T  2.0T  95% /mnt/user0
    /dev/loop2            270G   16G  255G   6% /var/lib/docker
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/ceaae5d21858eed21b3c02b988e671146808da6ac5659070f89479f8771981cf/merged
    /dev/loop3             10G  4.5M  9.5G   1% /etc/libvirt
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/1fe89828cb1992deac96043a165b0c2d257bee31fbdfba6cde3c12fa78108cbd/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/232e8202b591222a7e925454bd743b305391cf9b1ba5cf8e45f8776b0dd2d3c7/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/3f713ae84e4622598210b94b15793c3260aea4a692cc7d7226b55030c88d7a4f/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/7175475e611234ca5e3da0f392bc31f9db2ca0cddab1e36085a67b82094d12ee/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/39b7105f858ae7dfd92628e545a60bed09bf2fe79d22df47f2b715d566c0bbb1/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/9b4d538da85729710cdc271557e36a0118c0a00fa390174430df405af3f7c9e8/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/3bbc4e85d87c6c86fe05edc45febfd136d2e1cc08e61ea0e994b284db0fdc5c8/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/4ab32eea32431398b12171c4d5dc4b1f1e5284e29507fd487befacc6461e1750/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/71cbdeea3c3dd7e4ee0488597410796130d1a57a100cb50d67cf8365e89ba2a1/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/b9b2da94a87c92e6b24a4c5441d85b3b6b3a46a394f88a2fc7191ad27f1f875b/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/51e6dec694ebd54d74406ddfaecfb0e016166cc87f784b17f032982678de8700/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/b7bab57da934f09a97e8012dc9c48e9d2ac97a19edf5c6a22160b7c87b12eb00/merged
    overlay               270G   16G  255G   6% /var/lib/docker/overlay2/7ab691d92f572e5b56d725dad3662d095d7aaa71f9377b3eddd8a20f94d1e7f2/merged
    tmpfs                 3.2G     0  3.2G   0% /run/user/0

     

    idean

    Posted

    No docker, but I require NFS. 

     

    Was cleaning up some stuff using ncdu, deleted a folder, boom.

    I'm currently in the trial mode, now I'm rethinking this whole thing. Maybe I should stay with mergerfs+snapraid on Ubuntu...

     

    [Mon Feb 19 12:04:58 2024] ------------[ cut here ]------------
    [Mon Feb 19 12:04:58 2024] nfsd: non-standard errno: -103
    [Mon Feb 19 12:04:58 2024] WARNING: CPU: 1 PID: 24641 at fs/nfsd/nfsproc.c:909 nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024] Modules linked in: nfsv3 nfsv4 dns_resolver nfs ipmi_devintf rpcsec_gss_krb5 md_mod xfs dm_crypt dm_mod nfsd auth_rpcgss oid_registry lockd grace sunrpc zfs(PO) zunicode(PO) zzstd(O) zlua(O) zavl(PO) icp(PO) zcommon(PO) znvpair(PO) spl(O) xt_MASQUERADE xt_tcpudp xt_mark iptable_nat ip6table_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 tun tcp_diag inet_diag ip6table_filter ip6_tables iptable_filter ip_tables x_tables efivarfs af_packet 8021q garp mrp bridge stp llc bonding tls ixgbe xfrm_algo mdio intel_rapl_msr intel_rapl_common iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm ast drm_vram_helper i2c_algo_bit drm_ttm_helper ttm drm_kms_helper crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel drm sha512_ssse3 sha256_ssse3 sha1_ssse3 aesni_intel crypto_simd ipmi_ssif cryptd backlight rapl intel_cstate i2c_i801 agpgart qat_c3xxx i2c_smbus intel_qat i2c_core acpi_ipmi dh_generic rsa_generic syscopyarea input_leds ahci mpi sysfillrect
    [Mon Feb 19 12:04:58 2024]  sysimgblt crc8 libahci joydev led_class fb_sys_fops asn1_decoder ipmi_si button acpi_cpufreq unix [last unloaded: md_mod]
    [Mon Feb 19 12:04:58 2024] CPU: 1 PID: 24641 Comm: nfsd Tainted: P           O       6.1.64-Unraid #1
    [Mon Feb 19 12:04:58 2024] Hardware name: iXsystems FREENAS-MINI-3.0-XL+/A2SDi-H-TF, BIOS 1.1c 06/25/2019
    [Mon Feb 19 12:04:58 2024] RIP: 0010:nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024] Code: c3 cc cc cc cc 48 ff c0 48 83 f8 26 75 e0 80 3d dd c9 05 00 00 75 15 48 c7 c7 b5 52 dc a0 c6 05 cd c9 05 00 01 e8 01 a9 2d e0 <0f> 0b b8 00 00 00 05 c3 cc cc cc cc 48 83 ec 18 31 c9 ba ff 07 00
    [Mon Feb 19 12:04:58 2024] RSP: 0018:ffffc90001cd7d18 EFLAGS: 00010286
    [Mon Feb 19 12:04:58 2024] RAX: 0000000000000000 RBX: ffff888186d57000 RCX: 0000000000000027
    [Mon Feb 19 12:04:58 2024] RDX: 0000000000000002 RSI: ffffffff820d7e01 RDI: 00000000ffffffff
    [Mon Feb 19 12:04:58 2024] RBP: ffff888186d57180 R08: 0000000000000000 R09: ffffffff82245f10
    [Mon Feb 19 12:04:58 2024] R10: 00007fffffffffff R11: ffffffff82966af6 R12: 000000000000002e
    [Mon Feb 19 12:04:58 2024] R13: ffff88847ad070b8 R14: ffff88847ad070ec R15: 0000000000000026
    [Mon Feb 19 12:04:58 2024] FS:  0000000000000000(0000) GS:ffff88885fc40000(0000) knlGS:0000000000000000
    [Mon Feb 19 12:04:58 2024] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    [Mon Feb 19 12:04:58 2024] CR2: 0000145c47d4e690 CR3: 000000000420a000 CR4: 00000000003506e0
    [Mon Feb 19 12:04:58 2024] Call Trace:
    [Mon Feb 19 12:04:58 2024]  <TASK>
    [Mon Feb 19 12:04:58 2024]  ? __warn+0xab/0x122
    [Mon Feb 19 12:04:58 2024]  ? report_bug+0x109/0x17e
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? handle_bug+0x41/0x6f
    [Mon Feb 19 12:04:58 2024]  ? exc_invalid_op+0x13/0x60
    [Mon Feb 19 12:04:58 2024]  ? asm_exc_invalid_op+0x16/0x20
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd_rename+0x368/0x3d0 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd4_rename+0x61/0x8f [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd4_proc_compound+0x43f/0x575 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd_dispatch+0x1a9/0x262 [nfsd]
    [Mon Feb 19 12:04:58 2024]  svc_process_common+0x332/0x4df [sunrpc]
    [Mon Feb 19 12:04:58 2024]  ? ktime_get+0x35/0x49
    [Mon Feb 19 12:04:58 2024]  ? nfsd_svc+0x2b6/0x2b6 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? nfsd_shutdown_threads+0x5b/0x5b [nfsd]
    [Mon Feb 19 12:04:58 2024]  svc_process+0xc7/0xe4 [sunrpc]
    [Mon Feb 19 12:04:58 2024]  nfsd+0xd5/0x155 [nfsd]
    [Mon Feb 19 12:04:58 2024]  kthread+0xe7/0xef
    [Mon Feb 19 12:04:58 2024]  ? kthread_complete_and_exit+0x1b/0x1b
    [Mon Feb 19 12:04:58 2024]  ret_from_fork+0x22/0x30
    [Mon Feb 19 12:04:58 2024]  </TASK>
    [Mon Feb 19 12:04:58 2024] ---[ end trace 0000000000000000 ]---

     

    idean

    Posted

    14 minutes ago, idean said:

    No docker, but I require NFS. 

     

    Was cleaning up some stuff using ncdu, deleted a folder, boom.

    I'm currently in the trial mode, now I'm rethinking this whole thing. Maybe I should stay with mergerfs+snapraid on Ubuntu...

     

    [Mon Feb 19 12:04:58 2024] ------------[ cut here ]------------
    [Mon Feb 19 12:04:58 2024] nfsd: non-standard errno: -103
    [Mon Feb 19 12:04:58 2024] WARNING: CPU: 1 PID: 24641 at fs/nfsd/nfsproc.c:909 nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024] Modules linked in: nfsv3 nfsv4 dns_resolver nfs ipmi_devintf rpcsec_gss_krb5 md_mod xfs dm_crypt dm_mod nfsd auth_rpcgss oid_registry lockd grace sunrpc zfs(PO) zunicode(PO) zzstd(O) zlua(O) zavl(PO) icp(PO) zcommon(PO) znvpair(PO) spl(O) xt_MASQUERADE xt_tcpudp xt_mark iptable_nat ip6table_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 tun tcp_diag inet_diag ip6table_filter ip6_tables iptable_filter ip_tables x_tables efivarfs af_packet 8021q garp mrp bridge stp llc bonding tls ixgbe xfrm_algo mdio intel_rapl_msr intel_rapl_common iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm ast drm_vram_helper i2c_algo_bit drm_ttm_helper ttm drm_kms_helper crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel drm sha512_ssse3 sha256_ssse3 sha1_ssse3 aesni_intel crypto_simd ipmi_ssif cryptd backlight rapl intel_cstate i2c_i801 agpgart qat_c3xxx i2c_smbus intel_qat i2c_core acpi_ipmi dh_generic rsa_generic syscopyarea input_leds ahci mpi sysfillrect
    [Mon Feb 19 12:04:58 2024]  sysimgblt crc8 libahci joydev led_class fb_sys_fops asn1_decoder ipmi_si button acpi_cpufreq unix [last unloaded: md_mod]
    [Mon Feb 19 12:04:58 2024] CPU: 1 PID: 24641 Comm: nfsd Tainted: P           O       6.1.64-Unraid #1
    [Mon Feb 19 12:04:58 2024] Hardware name: iXsystems FREENAS-MINI-3.0-XL+/A2SDi-H-TF, BIOS 1.1c 06/25/2019
    [Mon Feb 19 12:04:58 2024] RIP: 0010:nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024] Code: c3 cc cc cc cc 48 ff c0 48 83 f8 26 75 e0 80 3d dd c9 05 00 00 75 15 48 c7 c7 b5 52 dc a0 c6 05 cd c9 05 00 01 e8 01 a9 2d e0 <0f> 0b b8 00 00 00 05 c3 cc cc cc cc 48 83 ec 18 31 c9 ba ff 07 00
    [Mon Feb 19 12:04:58 2024] RSP: 0018:ffffc90001cd7d18 EFLAGS: 00010286
    [Mon Feb 19 12:04:58 2024] RAX: 0000000000000000 RBX: ffff888186d57000 RCX: 0000000000000027
    [Mon Feb 19 12:04:58 2024] RDX: 0000000000000002 RSI: ffffffff820d7e01 RDI: 00000000ffffffff
    [Mon Feb 19 12:04:58 2024] RBP: ffff888186d57180 R08: 0000000000000000 R09: ffffffff82245f10
    [Mon Feb 19 12:04:58 2024] R10: 00007fffffffffff R11: ffffffff82966af6 R12: 000000000000002e
    [Mon Feb 19 12:04:58 2024] R13: ffff88847ad070b8 R14: ffff88847ad070ec R15: 0000000000000026
    [Mon Feb 19 12:04:58 2024] FS:  0000000000000000(0000) GS:ffff88885fc40000(0000) knlGS:0000000000000000
    [Mon Feb 19 12:04:58 2024] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    [Mon Feb 19 12:04:58 2024] CR2: 0000145c47d4e690 CR3: 000000000420a000 CR4: 00000000003506e0
    [Mon Feb 19 12:04:58 2024] Call Trace:
    [Mon Feb 19 12:04:58 2024]  <TASK>
    [Mon Feb 19 12:04:58 2024]  ? __warn+0xab/0x122
    [Mon Feb 19 12:04:58 2024]  ? report_bug+0x109/0x17e
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? handle_bug+0x41/0x6f
    [Mon Feb 19 12:04:58 2024]  ? exc_invalid_op+0x13/0x60
    [Mon Feb 19 12:04:58 2024]  ? asm_exc_invalid_op+0x16/0x20
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? nfserrno+0x45/0x51 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd_rename+0x368/0x3d0 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd4_rename+0x61/0x8f [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd4_proc_compound+0x43f/0x575 [nfsd]
    [Mon Feb 19 12:04:58 2024]  nfsd_dispatch+0x1a9/0x262 [nfsd]
    [Mon Feb 19 12:04:58 2024]  svc_process_common+0x332/0x4df [sunrpc]
    [Mon Feb 19 12:04:58 2024]  ? ktime_get+0x35/0x49
    [Mon Feb 19 12:04:58 2024]  ? nfsd_svc+0x2b6/0x2b6 [nfsd]
    [Mon Feb 19 12:04:58 2024]  ? nfsd_shutdown_threads+0x5b/0x5b [nfsd]
    [Mon Feb 19 12:04:58 2024]  svc_process+0xc7/0xe4 [sunrpc]
    [Mon Feb 19 12:04:58 2024]  nfsd+0xd5/0x155 [nfsd]
    [Mon Feb 19 12:04:58 2024]  kthread+0xe7/0xef
    [Mon Feb 19 12:04:58 2024]  ? kthread_complete_and_exit+0x1b/0x1b
    [Mon Feb 19 12:04:58 2024]  ret_from_fork+0x22/0x30
    [Mon Feb 19 12:04:58 2024]  </TASK>
    [Mon Feb 19 12:04:58 2024] ---[ end trace 0000000000000000 ]---

     

     

    unraid-diagnostics-20240219-1246.zip

    trurl

    Posted

    Are you sure you posted to the right thread? Diagnostics shows you have /mnt/user.

    dpunk3

    Posted

    This happens to me specifically when I use Filebot in a docker container to update filenames. Randomly, user folder will drop off entirely.

    Amane

    Posted

    Unfortunately, I also have to count myself among the victims...

    Sorry for creating a new bug report, I only found that thanks to @JorgeB..

    mbc0

    Posted

    This was fixed for me by moving from macvlan to ipvlan

    jaclas

    Posted (edited)

    I was working today on a virtual machine (running on my desktop) to which I have an NFS share mounted. Suddenly the share disappeared, the tools started showing errors, it surprised me. I went into Unraid and saw that all the shares were gone.
     

    image.thumb.png.f3005ab0016f47988dfa935a86bbc9fe.png

     

    Then I checked the mount points:

     

    image.png.9d9376c26387d7044d393ac48c5b8669.png

     

    I attach the diagnostics report.

     

    Then I started googling and found a lot of threads with this problem.
    I will add that I use NFS and will continue to use NFS, it is one of the key features for me, I will not give it up. Samba is on average 3 times slower (I did tests) and I'd rather change servers than give up NFS.

     

    I also came across an interesting thread on reddit, where someone stated that the problem is due to a very low limit of open files. That could be right, because in the background my server was processing thousands of mp3 files and I was programming in Node.js at the time, where a project in the ./node_modules subdirectory can also have tens of thousands of files.

     

     

     

    unraid-diagnostics-20240502-1108.zip

    Edited by jaclas
    robertklep

    Posted (edited)

    58 minutes ago, jaclas said:

    I will add that I use NFS and will continue to use NFS, it is one of the key features for me, I will not give it up. Samba is on average 3 times slower (I did tests) and I'd rather change servers than give up NFS.

    Switching to Samba and disabling NFS didn't fix the issue for me anyway, so it's not exclusive to using NFS. I moved away from Unraid to plain Ubuntu months ago, on the same hardware, and never had any issues since (uptimes of many months 🥳).

    Edited by robertklep
    primeval_god

    Posted (edited)

    On 5/2/2024 at 12:47 PM, jaclas said:

    I also came across an interesting thread on reddit, where someone stated that the problem is due to a very low limit of open files. That could be right, because in the background my server was processing thousands of mp3 files and I was programming in Node.js at the time, where a project in the ./node_modules subdirectory can also have tens of thousands of files.

    unraid-diagnostics-20240502-1108.zip

    If you happen to use Netdata for system monitoring you may be able to spot the circumstances leading up to the crash if the open file limit is the issue. Netdata has a section under Applications (apps.files) that shows the open files for different process groups. By default SHFS is not listed, but its open files are likely group under system or other. With a little customization of the netdata config you can make shfs show up as a separate entity. (Creating a custom apps_groups.conf file in the container and adding line like "shfs: shfs")

    Edited by primeval_god



    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Restore formatting

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Status Definitions

     

    Open = Under consideration.

     

    Solved = The issue has been resolved.

     

    Solved version = The issue has been resolved in the indicated release version.

     

    Closed = Feedback or opinion better posted on our forum for discussion. Also for reports we cannot reproduce or need more information. In this case just add a comment and we will review it again.

     

    Retest = Please retest in latest release.


    Priority Definitions

     

    Minor = Something not working correctly.

     

    Urgent = Server crash, data loss, or other showstopper.

     

    Annoyance = Doesn't affect functionality but should be fixed.

     

    Other = Announcement or other non-issue.

×
×
  • Create New...