• [6.6.0] NFS Kernel crash


    ajeffco
    • Solved Urgent

    Hello,

     

    Running unraid 6.6.0 stable, with a mostly NFS shares server.  NFS appears to be crashing.  Below is the first indication in the log file of a problem.  All clients lock up with "nfs: server tower not responding, timed out" from that point forward.  I have a coworker running unraid who has had the same issue, and while we initially thought it was just NFS, all CIFS AND RSYNC shares become unavailable also when this happens.  When this happens unraid becomes 100% unusable for file operations for any client!

     

    This appears to  have been reported already at [ 6.6.0-RC4 ] NFS CRASHES.  I submitted another since this is 6.6.0 stable.

     

    HOW TO REPRODUCE:  Reboot and just wait.  My coworker has had this happen a few times, this is my first issue.

     

    Sep 26 03:48:41 tower kernel: ------------[ cut here ]------------
    Sep 26 03:48:41 tower kernel: nfsd: non-standard errno: -103
    Sep 26 03:48:41 tower kernel: WARNING: CPU: 2 PID: 12478 at fs/nfsd/nfsproc.c:817 nfserrno+0x44/0x4a [nfsd]
    Sep 26 03:48:41 tower kernel: Modules linked in: md_mod nfsd lockd grace sunrpc bonding mlx4_en mlx4_core igb sb_edac x86_pkg_temp_thermal intel_powerclamp coretemp ast ttm kvm_intel drm_kms_helper kvm drm crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel pcbc aesni_intel aes_x86_64 crypto_simd cryptd agpgart glue_helper intel_cstate intel_uncore ipmi_ssif intel_rapl_perf syscopyarea mpt3sas i2c_i801 i2c_algo_bit i2c_core ahci sysfillrect pcc_cpufreq libahci sysimgblt fb_sys_fops raid_class scsi_transport_sas wmi acpi_power_meter ipmi_si acpi_pad button [last unloaded: md_mod]
    Sep 26 03:48:41 tower kernel: CPU: 2 PID: 12478 Comm: nfsd Not tainted 4.18.8-unRAID #1
    Sep 26 03:48:41 tower kernel: Hardware name: Supermicro Super Server/X10SRL-F, BIOS 3.0a 02/08/2018
    Sep 26 03:48:41 tower kernel: RIP: 0010:nfserrno+0x44/0x4a [nfsd]
    Sep 26 03:48:41 tower kernel: Code: c0 48 83 f8 22 75 e2 80 3d b3 06 01 00 00 bb 00 00 00 05 75 17 89 fe 48 c7 c7 3b 9a 18 a0 c6 05 9c 06 01 00 01 e8 8a ec ec e0 <0f> 0b 89 d8 5b c3 48 83 ec 18 31 c9 ba ff 07 00 00 65 48 8b 04 25 
    Sep 26 03:48:41 tower kernel: RSP: 0018:ffffc9000c743db8 EFLAGS: 00010286
    Sep 26 03:48:41 tower kernel: RAX: 0000000000000000 RBX: 0000000005000000 RCX: 0000000000000007
    Sep 26 03:48:41 tower kernel: RDX: 0000000000000000 RSI: ffff88087fc96470 RDI: ffff88087fc96470
    Sep 26 03:48:41 tower kernel: RBP: ffffc9000c743e08 R08: 0000000000000003 R09: ffffffff82202400
    Sep 26 03:48:41 tower kernel: R10: 000000000000087f R11: 000000000000a9e4 R12: ffff8802b01ea808
    Sep 26 03:48:41 tower kernel: R13: ffff8807febb2a58 R14: 0000000000000002 R15: ffffffffa01892a0
    Sep 26 03:48:41 tower kernel: FS:  0000000000000000(0000) GS:ffff88087fc80000(0000) knlGS:0000000000000000
    Sep 26 03:48:41 tower kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    Sep 26 03:48:41 tower kernel: CR2: 00001501e0097000 CR3: 0000000001e0a005 CR4: 00000000001606e0
    Sep 26 03:48:41 tower kernel: Call Trace:
    Sep 26 03:48:41 tower kernel: nfsd_open+0x15e/0x17c [nfsd]
    Sep 26 03:48:41 tower kernel: nfsd_write+0x4c/0xaa [nfsd]
    Sep 26 03:48:41 tower kernel: nfsd3_proc_write+0xad/0xdb [nfsd]
    Sep 26 03:48:41 tower kernel: nfsd_dispatch+0xb4/0x169 [nfsd]
    Sep 26 03:48:41 tower kernel: svc_process+0x4b5/0x666 [sunrpc]
    Sep 26 03:48:41 tower kernel: ? nfsd_destroy+0x48/0x48 [nfsd]
    Sep 26 03:48:41 tower kernel: nfsd+0xeb/0x142 [nfsd]
    Sep 26 03:48:41 tower kernel: kthread+0x10b/0x113
    Sep 26 03:48:41 tower kernel: ? kthread_flush_work_fn+0x9/0x9
    Sep 26 03:48:41 tower kernel: ret_from_fork+0x35/0x40
    Sep 26 03:48:41 tower kernel: ---[ end trace 0df913a547279c0d ]---

    tower-diagnostics-20180926-0904.zip




    User Feedback

    Recommended Comments



    15 hours ago, ajeffco said:

    Hello,

     

    Another night of no crash.

    
    root@tower:~# uptime
     11:28:15 up 1 day, 12:58,  1 user,  load average: 0.01, 0.00, 0.00

     

    @Frank76 Sorry to hear your still having problems.  I've run Synology backups manually and let them run by schedule and haven't had trouble since the two changes.  I've also run 2 macbook timemachine backups at the same time as a manual synology backup scan each day since the changes, and it hasn't crashed.  I want to say that for certain one of my crashes occurred when there was no I/O going to the unraid rig.

    I just did another test, and it took a whopping 17 minutes before the storage locked up again. I'm using Cloudberry from a vm, with the data source being a nfs mounted unraid share, and the destination is an automounted sshfs filesystem running to a disk at a friend's house. It has been working well for a few months now until recently. I'm also wondering if the new feature block level backup feature of cloudberry that could be compounding the issue. I'll disable the block level backup and see if I can get a successful backup.

    Link to comment

    Having the same issue with NFS on my Dell T130 server, forced me to downgrade after the first NFS crash.  Seems like core functionality testing was skipped in 6.6 release as NFS/SMB are core components of a NAS not fluff like docker and kvm or a GUI that looks pretty on a cell phone. 

    Link to comment
    23 minutes ago, klamath said:

    Having the same issue with NFS on my Dell T130 server, forced me to downgrade after the first NFS crash.  Seems like core functionality testing was skipped in 6.6 release as NFS/SMB are core components of a NAS not fluff like docker and kvm or a GUI that looks pretty on a cell phone. 

    Did you try 6.6.1?

    I have had no crashes since trying that so far.

    I use NFS share for my home media center.

    Link to comment
    2 hours ago, nekromantik said:

    Did you try 6.6.1?

    I have had no crashes since trying that so far.

    I use NFS share for my home media center.

    I have not, once this thread reports back A-OK I will upgrade.

    Link to comment

    got 11hrs this time before it crashed again...i went back to stability with 6.5.3 once again

     

    @limetech NFS is still crashing.....lasted 11hrs this time this is killing the user folder for some reason....i hope you and the team can find out whats causing this and get it fixed ASAP its very annoying


    Sep 30 23:23:56 Thor kernel: ------------[ cut here ]------------
    Sep 30 23:23:56 Thor kernel: nfsd: non-standard errno: -103
    Sep 30 23:23:56 Thor kernel: WARNING: CPU: 1 PID: 6507 at fs/nfsd/nfsproc.c:817 nfserrno+0x44/0x4a [nfsd]
    Sep 30 23:23:56 Thor kernel: Modules linked in: xt_nat macvlan xt_CHECKSUM iptable_mangle ipt_REJECT ebtable_filter ebtables ip6table_filter ip6_tables vhost_net tun vhost tap ipt_MASQUERADE iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 iptable_filter ip_tables nf_nat arc4 ecb md4 sha512_ssse3 sha512_generic cmac cifs ccm xfs nfsd lockd grace sunrpc md_mod i915 i2c_algo_bit iosf_mbi drm_kms_helper drm intel_gtt agpgart syscopyarea sysfillrect sysimgblt fb_sys_fops x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel pcbc aesni_intel aes_x86_64 crypto_simd cryptd mpt3sas glue_helper alx intel_cstate intel_uncore intel_rapl_perf ahci raid_class libahci scsi_transport_sas mxm_wmi i2c_i801 mdio i2c_core video wmi mlx4_core backlight thermal
    Sep 30 23:23:56 Thor kernel: button acpi_pad pcc_cpufreq fan
    Sep 30 23:23:56 Thor kernel: CPU: 1 PID: 6507 Comm: nfsd Not tainted 4.18.10-unRAID #2
    Sep 30 23:23:56 Thor kernel: Hardware name: MSI MS-7978/H170 GAMING M3 (MS-7978), BIOS C.B0 01/25/2018
    Sep 30 23:23:56 Thor kernel: RIP: 0010:nfserrno+0x44/0x4a [nfsd]
    Sep 30 23:23:56 Thor kernel: Code: c0 48 83 f8 22 75 e2 80 3d b3 06 01 00 00 bb 00 00 00 05 75 17 89 fe 48 c7 c7 3b 9a 47 a0 c6 05 9c 06 01 00 01 e8 3b ed bd e0 <0f> 0b 89 d8 5b c3 48 83 ec 18 31 c9 ba ff 07 00 00 65 48 8b 04 25 
    Sep 30 23:23:56 Thor kernel: RSP: 0018:ffffc9000377bdc0 EFLAGS: 00010282
    Sep 30 23:23:56 Thor kernel: RAX: 0000000000000000 RBX: 0000000005000000 RCX: 0000000000000007
    Sep 30 23:23:56 Thor kernel: RDX: 0000000000000000 RSI: ffff88086dc56470 RDI: ffff88086dc56470
    Sep 30 23:23:56 Thor kernel: RBP: ffffc9000377be10 R08: 0000000000000003 R09: ffffffff8220d400
    Sep 30 23:23:56 Thor kernel: R10: 0000000000000445 R11: 0000000000015a7c R12: ffff880845a49c08
    Sep 30 23:23:56 Thor kernel: R13: 000000001373b000 R14: ffff880845a49d68 R15: 0000000000000005
    Sep 30 23:23:56 Thor kernel: FS: 0000000000000000(0000) GS:ffff88086dc40000(0000) knlGS:0000000000000000
    Sep 30 23:23:56 Thor kernel: CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    Sep 30 23:23:56 Thor kernel: CR2: 000055d8c34aa2f8 CR3: 0000000001e0a002 CR4: 00000000003626e0
    Sep 30 23:23:56 Thor kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
    Sep 30 23:23:56 Thor kernel: DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
    Sep 30 23:23:56 Thor kernel: Call Trace:
    Sep 30 23:23:56 Thor kernel: nfsd_open+0x15e/0x17c [nfsd]
    Sep 30 23:23:56 Thor kernel: nfsd_read+0x45/0xec [nfsd]
    Sep 30 23:23:56 Thor kernel: nfsd3_proc_read+0x95/0xda [nfsd]
    Sep 30 23:23:56 Thor kernel: nfsd_dispatch+0xb4/0x169 [nfsd]
    Sep 30 23:23:56 Thor kernel: svc_process+0x4b5/0x666 [sunrpc]
    Sep 30 23:23:56 Thor kernel: ? nfsd_destroy+0x48/0x48 [nfsd]
    Sep 30 23:23:56 Thor kernel: nfsd+0xeb/0x142 [nfsd]
    Sep 30 23:23:56 Thor kernel: kthread+0x10b/0x113
    Sep 30 23:23:56 Thor kernel: ? kthread_flush_work_fn+0x9/0x9
    Sep 30 23:23:56 Thor kernel: ret_from_fork+0x35/0x40
    Sep 30 23:23:56 Thor kernel: ---[ end trace cd5759a6b02d0b74 ]---
    Sep 30 23:44:16 Thor sshd[22182]: SSH: Server;Ltype: Version;Remote: 10.0.0.10-65250;Protocol: 2.0;Client: OpenSSH_7.7
    Sep 30 23:44:16 Thor sshd[22182]: SSH: Server;Ltype: Kex;Remote: 10.0.0.10-65250;Enc: [email protected];MAC: <implicit>;Comp: none [preauth]
    Sep 30 23:44:16 Thor sshd[22182]: SSH: Server;Ltype: Authname;Remote: 10.0.0.10-65250;Name: root [preauth]
    Sep 30 23:44:20 Thor sshd[22182]: Accepted password for root from 10.0.0.10 port 65250 ssh2
    Sep 30 23:46:53 Thor emhttpd: req (7): startState=STARTED&cmdStop=Stop&csrf_token=****************&csrf_token=****************
    Sep 30 23:46:53 Thor emhttpd: Spinning up all drives...
    Sep 30 23:46:53 Thor emhttpd: shcmd (151): /usr/sbin/hdparm -S0 /dev/sdo
    Sep 30 23:46:53 Thor kernel: mdcmd (48): nocheck 
    Sep 30 23:46:53 Thor kernel: md: nocheck_array: check not active
    Sep 30 23:46:53 Thor kernel: mdcmd (49): spinup 0
    Sep 30 23:46:53 Thor kernel: mdcmd (50): spinup 1
    Sep 30 23:46:53 Thor kernel: mdcmd (51): spinup 2
    Sep 30 23:46:53 Thor kernel: mdcmd (52): spinup 3
    Sep 30 23:46:53 Thor kernel: mdcmd (53): spinup 4
    Sep 30 23:46:53 Thor kernel: mdcmd (54): spinup 5
    Sep 30 23:46:53 Thor kernel: mdcmd (55): spinup 6
    Sep 30 23:46:53 Thor kernel: mdcmd (56): spinup 7
    Sep 30 23:46:53 Thor kernel: mdcmd (57): spinup 8
    Sep 30 23:46:53 Thor kernel: mdcmd (58): spinup 9
    Sep 30 23:46:53 Thor kernel: mdcmd (59): spinup 10
    Sep 30 23:46:53 Thor kernel: mdcmd (60): spinup 29
    Sep 30 23:46:53 Thor root: 
    Sep 30 23:46:53 Thor root: /dev/sdo:
    Sep 30 23:46:53 Thor root: setting standby to 0 (off)
    Sep 30 23:46:53 Thor emhttpd: shcmd (152): /usr/sbin/hdparm -S0 /dev/sdn
    Sep 30 23:46:53 Thor root: 
    Sep 30 23:46:53 Thor root: /dev/sdn:
     

    Link to comment
    4 hours ago, Can0nfan said:

    @Frank76 downgrading is pretty easy if you need a copy of the 6.5.3 zip file it should still be available under downloads but if not I have a copy for this very reason. DM me if you need it

    https://s3.amazonaws.com/dnld.lime-technology.com/stable/unRAIDServer-6.5.3-x86_64.zip

    Thanks for the link. I thought I was on to something because once I disabled the block level backups, it was looking like it was working. It actually finished a few jobs. but sadly it just died on me again after less than 24 hours. I will downgrade as soon as I can. I'm not really clear on the procedure to downgrade. Do I extract the zip file and overwrite what is in the /boot/previous directory, then go into the webui, tools, update os and select previous version?

     

    Thanks!

    Link to comment
    3 minutes ago, Frank76 said:

    Thanks for the link. I thought I was on to something because once I disabled the block level backups, it was looking like it was working. It actually finished a few jobs. but sadly it just died on me again after less than 24 hours. I will downgrade as soon as I can. I'm not really clear on the procedure to downgrade. Do I extract the zip file and overwrite what is in the /boot/previous directory, then go into the webui, tools, update os and select previous version?

     

    Thanks!

    you extract it and just replace the bz* files from the 6.5.3 zip on the flash drive and reboot..

     

    mine too lasted 11hrs before the crash of NFS again im back on the more stable 6.5.3 until LimeTech can get a handle on this issue

    Edited by Can0nfan
    Link to comment
    root@tower:~# uptime
     08:09:56 up 3 days,  9:40,  1 user,  load average: 2.50, 2.59, 2.15

    No trouble since the 2 changes. 

    Edited by ajeffco
    Link to comment

    Yeah touch wood 3 days no crashes.

    I think the change in 6.6.1 only seemed to help those who dont use backups.

    Mine is just used as a NFS NAS for my Media Center and also as a docker server.

    Link to comment

    I'm using it for backups also.

     

    Time Machine from 2 macbook pros

    Synology Hyberbackup

    Proxmox VM Backups from 4 nodes, 13 VM's

     

    In addition to the media center stuff (including the sab download target).

     

    I'm using cache disks on my array, wonder if that makes a difference.

    Link to comment

    Limetech has found this is related to FUSER crashing which crashes the user folder which crashes NFS and is looking into which kernel the issue started happening on

    Edited by Can0nfan
    Link to comment
    On 9/30/2018 at 10:34 AM, nekromantik said:

    Did you try 6.6.1?

    I have had no crashes since trying that so far.

    I use NFS share for my home media center.

    I submitted a feature request here for an alternative NFS server in unraid.

      

    Link to comment

    While I know some people appear to still be having trouble with it, I wanted to give feedback that since the 6.6.1 upgrade I have experienced no crashes.

     

    root@tower:~# uptime
     00:54:53 up 10 days,  2:25,  1 user,  load average: 0.01, 0.11, 0.24
    root@tower:~#

     

    Link to comment

    I'm having the same issues.  I'm exporting most of my shares via CIFS and NFS.  I have an Opensuse 42.3 VM mounting most of the exports, including for /home.

     

    (Lots of the following in syslog)
    Oct 10 20:53:23 samurai rpcbind[3666]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:54:24 samurai rpcbind[4265]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:55:29 samurai rpcbind[4894]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:56:30 samurai rpcbind[5473]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:57:32 samurai rpcbind[6093]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:58:35 samurai rpcbind[6816]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 20:59:37 samurai rpcbind[7511]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 21:00:49 samurai rpcbind[8309]: connect from 192.168.1.19 to getport/addr(nlockmgr)
    Oct 10 21:01:52 samurai rpcbind[8975]: connect from 192.168.1.19 to getport/addr(nlockmgr)
     

    Followed by:

     

    Oct 10 21:01:54 samurai kernel: ------------[ cut here ]------------
    Oct 10 21:01:54 samurai kernel: nfsd: non-standard errno: -103
    Oct 10 21:01:54 samurai kernel: WARNING: CPU: 4 PID: 4716 at fs/nfsd/nfsproc.c:817 nfserrno+0x44/0x4a [nfsd]
    Oct 10 21:01:54 samurai kernel: Modules linked in: xt_nat veth xt_CHECKSUM iptable_mangle ipt_REJECT ebtable_filter ebtables ip6table_filter ip6_tables vhost_net tun vhost tap ipt_MASQUERADE iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 iptable_filter ip_tables nf_nat nfsd lockd grace sunrpc md_mod bonding sr_mod cdrom x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel pcbc aesni_intel mpt3sas aes_x86_64 crypto_simd cryptd glue_helper raid_class scsi_transport_sas intel_cstate intel_uncore ahci libahci i2c_i801 intel_rapl_perf i2c_core e1000e video mxm_wmi wmi backlight pcc_cpufreq thermal acpi_pad button fan
    Oct 10 21:01:54 samurai kernel: CPU: 4 PID: 4716 Comm: nfsd Not tainted 4.18.10-unRAID #2
    Oct 10 21:01:54 samurai kernel: Hardware name: Gigabyte Technology Co., Ltd. Z97X-UD3H/Z97X-UD3H-CF, BIOS F10b 03/03/2016
    Oct 10 21:01:54 samurai kernel: RIP: 0010:nfserrno+0x44/0x4a [nfsd]
    Oct 10 21:01:54 samurai kernel: Code: c0 48 83 f8 22 75 e2 80 3d b3 06 01 00 00 bb 00 00 00 05 75 17 89 fe 48 c7 c7 3b 7a 25 a0 c6 05 9c 06 01 00 01 e8 3b 0d e0 e0 <0f> 0b 89 d8 5b c3 48 83 ec 18 31 c9 ba ff 07 00 00 65 48 8b 04 25
    Oct 10 21:01:54 samurai kernel: RSP: 0018:ffffc90001d43dc0 EFLAGS: 00010282
    Oct 10 21:01:54 samurai kernel: RAX: 0000000000000000 RBX: 0000000005000000 RCX: 0000000000000007
    Oct 10 21:01:54 samurai kernel: RDX: 0000000000000000 RSI: ffff88042fb16470 RDI: ffff88042fb16470
    Oct 10 21:01:54 samurai kernel: RBP: ffffc90001d43e10 R08: 0000000000000003 R09: ffffffff8220db00
    Oct 10 21:01:54 samurai kernel: R10: 0000000000000458 R11: 00000000000160c8 R12: ffff88041a9b2408
    Oct 10 21:01:54 samurai kernel: R13: 0000000000008000 R14: ffff88041a9b2568 R15: 0000000000000008
    Oct 10 21:01:54 samurai kernel: FS:  0000000000000000(0000) GS:ffff88042fb00000(0000) knlGS:0000000000000000
    Oct 10 21:01:54 samurai kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
    Oct 10 21:01:54 samurai kernel: CR2: 000014bf59f56020 CR3: 0000000001e0a005 CR4: 00000000001626e0
    Oct 10 21:01:54 samurai kernel: Call Trace:
    Oct 10 21:01:54 samurai kernel: nfsd_open+0x15e/0x17c [nfsd]
    Oct 10 21:01:54 samurai kernel: nfsd_read+0x45/0xec [nfsd]
    Oct 10 21:01:54 samurai kernel: nfsd3_proc_read+0x95/0xda [nfsd]
    Oct 10 21:01:54 samurai kernel: nfsd_dispatch+0xb4/0x169 [nfsd]
    Oct 10 21:01:54 samurai kernel: svc_process+0x4b5/0x666 [sunrpc]
    Oct 10 21:01:54 samurai kernel: ? nfsd_destroy+0x48/0x48 [nfsd]
    Oct 10 21:01:54 samurai kernel: nfsd+0xeb/0x142 [nfsd]
    Oct 10 21:01:54 samurai kernel: kthread+0x10b/0x113
    Oct 10 21:01:54 samurai kernel: ? kthread_flush_work_fn+0x9/0x9
    Oct 10 21:01:54 samurai kernel: ret_from_fork+0x35/0x40
    Oct 10 21:01:54 samurai kernel: ---[ end trace 82cc1d618070c378 ]---
    Oct 10 21:02:06 samurai rpc.mountd[4719]: Cannot export /mnt/user/Projects, possibly unsupported filesystem or fsid= required
     

     

    I've attached diags.  Everything is remounted on /mnt/user0, but of course my NFS exports and CIFS shares aren't working.  I'm about to reboot the server. 

     

    What's the consensus solution?  Downgrade to 6.5.3 or what?  I upgraded to 6.6.1 because I was getting whole-server lockups & spontaneous reboots every couple of days all of a sudden, whereas 6.6.1 *seemed* more stable.

    unraid.diags.Oct11_2018.7z

    samurai-diagnostics-20181011-1547.zip

    Link to comment

    Kernel updated but nothing obviously patched that would explain this.  We did add more instrumentation into FUSE, would appreciate a retest and repost of diagnostics.zip upon failure.

    Link to comment

    Just crashed again on me with same results - updating to 6.6.2 and will upload diags when it happens again.

     

    FYI, what seems like a surefire way of killing the nfsserver and then the shfs (and making everything remount on /mnt/user0) is mounting one of the NFS exports as /home for a linux VM, especially with lots of little NFS I/O calls and NFS locks for Firefox, etc.  It gradually slows down more and more, until finally nfsd on unraid just croaks.

    Edited by Kamikazejs
    • Like 1
    • Upvote 1
    Link to comment
    On 10/14/2018 at 12:40 AM, limetech said:

    Kernel updated but nothing obviously patched that would explain this.  We did add more instrumentation into FUSE, would appreciate a retest and repost of diagnostics.zip upon failure.

    I'm at 16 hours uptime on 6.6.2 without a NFS crash or permission issues on my user share. so far so good for me...I will keep a close eye

    Edited by Can0nfan
    updated uptime
    Link to comment

    and food for thought......31 hours and counting, a alot of NFS use in this time frame with 0 issues. thanks @limetech 6.6.2 is the best so far of the 6.6.x releases!

     

    Link to comment

    I haven't had an issue since updating 2 days 6 hours ago. Looks like the issue has been resolved. Thank you very much for the quick resolution!

    Link to comment



    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Restore formatting

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Status Definitions

     

    Open = Under consideration.

     

    Solved = The issue has been resolved.

     

    Solved version = The issue has been resolved in the indicated release version.

     

    Closed = Feedback or opinion better posted on our forum for discussion. Also for reports we cannot reproduce or need more information. In this case just add a comment and we will review it again.

     

    Retest = Please retest in latest release.


    Priority Definitions

     

    Minor = Something not working correctly.

     

    Urgent = Server crash, data loss, or other showstopper.

     

    Annoyance = Doesn't affect functionality but should be fixed.

     

    Other = Announcement or other non-issue.