ndetar

Members
  • Posts

    41
  • Joined

  • Last visited

Posts posted by ndetar

  1. 31 minutes ago, MarbleComa said:

    Hi @ndetar, thanks for creating this unraid template! I've been trying for a few hours to get the deepstack gpu docker running on my system, but it seems that attempting to use a 3060 ti leads to quite some issues. With the default installation directly from the deepquestai/deepstack:gpu repository, nothing works, and I was able to determine that the included version of pytorch does not support the 3060 ti. I ran cat ../logs/stderr.txt and found the following message:

     

    "NVIDIA GeForce RTX 3060 ti with CUDA capability sm_86 is not compatible with the current PyTorch installation. The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_70.
    If you want to use the NVIDIA GeForce RTX 3060 ti GPU with PyTorch."

     

    I did some reading on the pytorch site (https://pytorch.org/get-started/locally/), and I found the pip install command to get their LTS (1.8.2) version which supports cuda 11.1. As a side note, I'm on Unraid 6.10.0-rc2 and the Nvidia driver is CUDA version 11.5. I updated both torch and torch vision in the container and that allows the object detection to run on the 3060 ti (some progress, yay!). Unfortunately, now I'm getting a new error with the face detection that I haven't been able to solve despite my best googling. The output of cat ../logs/stderr.txt is below:

     

        Traceback (most recent call last):
      File "/app/intelligencelayer/shared/face.py", line 307, in face
        det = detector.predict(img, 0.55)
      File "/app/intelligencelayer/shared/./process.py", line 61, in predict
        pred = self.model(img, augment=False)[0]
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1102, in _call_impl
        return forward_call(*input, **kwargs)
      File "/app/intelligencelayer/shared/./models/yolo.py", line 149, in forward
        return self.forward_once(x, profile)  # single-scale inference, train
      File "/app/intelligencelayer/shared/./models/yolo.py", line 176, in forward_once
        x = m(x)  # run
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1102, in _call_impl
        return forward_call(*input, **kwargs)
      File "/app/intelligencelayer/shared/./models/common.py", line 109, in forward
        1,
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1102, in _call_impl
        return forward_call(*input, **kwargs)
      File "/app/intelligencelayer/shared/./models/common.py", line 32, in forward
        return self.act(self.bn(self.conv(x)))
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1102, in _call_impl
        return forward_call(*input, **kwargs)
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/activation.py", line 461, in forward
        return F.hardswish(input, self.inplace)
      File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1178, in __getattr__
        type(self).__name__, name))
    AttributeError: 'Hardswish' object has no attribute 'inplace'

     

    Do you have any recommendation on how to proceed here? Many thanks!

    Further modifications are beyond what I am able to help with. You may have better luck posting or searching on the Deepstack forum. Sorry I can't be of much help. From what I can see it seems like the face recognition in Deepstack doesn't support the newer version of pytorch.

  2. With only one did the logs output the same errors?

    In extra parameters did you add: --runtime=nvidia

     

    Try removing the container image completely and pulling it again, and try the template I made in the app store fallowing the instructions for GPU use.

    I'm curious what the logs show when you use the template.

    image.png.e46892c325f4ef08003b9444239ce154.png

    • Like 1
  3. So if I understand the logs correctly I think that your GPU doesn't have enough vRAM for all the recognition processes you are trying to run. For me, just object detection takes half a gig of vRAM. Try running deepstack in GPU mode with fewer recognition APIs enable. Maybe just try object recognition by its self first and see if it runs.

  4. I have been using it with a GPU for a while now and its been working great. Could you provide some additional information such as the log output from the container, maybe a screenshot of your config, etc. It's hard to troubleshoot without more context.

  5. I recently ran into an issue I need help troubleshooting. I originally posted in the general support (HERE) but doing some troubleshooting myself I realized it was wireguard causing the issue. I had wireguard setup for remote LAN access only. 

     

    To summarize when I restart the docker service from the WebUI while wireguard was running all networking on Unraid would break. I wouldn't be able to ping the server from a separate PC and when I tried to ping anything from Unraid it was unable to reach anything. Everything would work fine as long as I didn't try to stop the docker service. I run a windows 10 VM with a GPU passed through and the VM would continue to function normally, it would even still have internet access, but was also not able to to access Unraid in any way.

    Troubleshooting was kind of a pain because I run Unraid headless (primary GPU is passed through to VM) so when the networking would go down I was unable to see the console (I could plugin a keyboard and blindly run commands, which is how I was able to get diagnostics and cleanly shutdown).

     

    I posted the log in my original post. Any help would be appreciated since I would like to have wireguard setup.

  6. After removing wireguard plugin and attempting to restart the docker service it work as it should. I suspect it has something to do with the routing rules wireguard sets up. I don't know why specifically stopping the docker service would make networking completely fall apart, but I'll reach out to the wireguard support specifically to see if I can further troubleshoot this. Marking this a solved.

  7. I restored from a backup of the flash drive taken overnight and was able to get the server up and running again, but attempting to stop the docker service caused the same thing to happen again. Everything works perfectly normal until I attempt to stop the service. The only thing I can think of that I changed recently was some changes I made in wireguard, so I don't know if that has something to do with this or what.

  8. After rebooting the networking seems to still be broken. I had ping running and it just hangs after a few pings and then there is nothing after that.

    From 192.168.1.162 icmp_seq=892 Destination Host Unreachable
    From 192.168.1.162 icmp_seq=895 Destination Host Unreachable
    From 192.168.1.162 icmp_seq=898 Destination Host Unreachable
    From 192.168.1.162 icmp_seq=901 Destination Host Unreachable
    From 192.168.1.162 icmp_seq=904 Destination Host Unreachable
    64 bytes from 192.168.1.6: icmp_seq=907 ttl=63 time=193 ms
    64 bytes from 192.168.1.6: icmp_seq=908 ttl=63 time=7.42 ms
    64 bytes from 192.168.1.6: icmp_seq=909 ttl=63 time=2.95 ms
    64 bytes from 192.168.1.6: icmp_seq=910 ttl=63 time=16.3 ms
    64 bytes from 192.168.1.6: icmp_seq=911 ttl=63 time=16.9 ms
    64 bytes from 192.168.1.6: icmp_seq=912 ttl=63 time=3.20 ms
    64 bytes from 192.168.1.6: icmp_seq=913 ttl=63 time=17.0 ms
    64 bytes from 192.168.1.6: icmp_seq=914 ttl=63 time=4.74 ms
    

     

     

    It wasn't until after I shutdown the server again that the Unreachable message started again.

  9. When I stop the docker service from the WebUI all networking seems to break. The WebUI stops loading, and when I ping the server from another PC it just hangs. From the server itself I can't ping anything. The Windows 10 VM I have running with GPU passthrough continues to function normally, networking even still works so I'm kind of at a lost as to what is going on. 

     

    I'm running the latest version of Unraid but with the AMD vender-reset added.

     

    This is everything in the log after I logged in and tried to stop the docker service.

     

    May  4 08:13:56 nateserver webGUI: Successful login user root from 192.168.1.49
    May  4 08:18:31 nateserver emhttpd: cmd: /usr/local/emhttp/plugins/dynamix/scripts/share_size Downloads ssz1 cache
    May  4 08:19:53 nateserver ool www[30048]: /usr/local/emhttp/plugins/dynamix/scripts/emcmd 'cmdStatus=Apply'
    May  4 08:19:53 nateserver emhttpd: Starting services...
    May  4 08:19:53 nateserver emhttpd: shcmd (158): /etc/rc.d/rc.samba restart
    May  4 08:19:53 nateserver winbindd[9396]: [2021/05/04 08:19:53.516830,  0] ../../source3/winbindd/winbindd.c:244(winbindd_sig_term_handler)
    May  4 08:19:53 nateserver winbindd[9396]:   Got sig[15] terminate (is_parent=0)
    May  4 08:19:53 nateserver winbindd[9980]: [2021/05/04 08:19:53.517005,  0] ../../source3/winbindd/winbindd.c:244(winbindd_sig_term_handler)
    May  4 08:19:53 nateserver winbindd[9980]:   Got sig[15] terminate (is_parent=0)
    May  4 08:19:53 nateserver wsdd[9975]: set_multicast: Failed to set IPv4 multicast
    ### [PREVIOUS LINE REPEATED 1 TIMES] ###
    May  4 08:19:53 nateserver nmbd[9968]: [2021/05/04 08:19:53.518043,  0] ../../source3/nmbd/nmbd.c:59(terminate)
    May  4 08:19:53 nateserver nmbd[9968]:   Got SIGTERM: going down...
    May  4 08:19:53 nateserver winbindd[9978]: [2021/05/04 08:19:53.518064,  0] ../../source3/winbindd/winbindd.c:244(winbindd_sig_term_handler)
    May  4 08:19:53 nateserver winbindd[9978]:   Got sig[15] terminate (is_parent=1)
    May  4 08:19:53 nateserver winbindd[11317]: [2021/05/04 08:19:53.518598,  0] ../../source3/winbindd/winbindd.c:244(winbindd_sig_term_handler)
    May  4 08:19:53 nateserver winbindd[11317]:   Got sig[15] terminate (is_parent=0)
    May  4 08:19:55 nateserver root: Starting Samba:  /usr/sbin/smbd -D
    May  4 08:19:55 nateserver root:                  /usr/sbin/nmbd -D
    May  4 08:19:55 nateserver smbd[3839]: [2021/05/04 08:19:55.671702,  0] ../../lib/util/become_daemon.c:135(daemon_ready)
    May  4 08:19:55 nateserver smbd[3839]:   daemon_ready: daemon 'smbd' finished starting up and ready to serve connections
    May  4 08:19:55 nateserver root:                  /usr/sbin/wsdd 
    May  4 08:19:55 nateserver nmbd[3844]: [2021/05/04 08:19:55.681068,  0] ../../lib/util/become_daemon.c:135(daemon_ready)
    May  4 08:19:55 nateserver nmbd[3844]:   daemon_ready: daemon 'nmbd' finished starting up and ready to serve connections
    May  4 08:19:55 nateserver root:                  /usr/sbin/winbindd -D
    May  4 08:19:55 nateserver wsdd[3851]: set_multicast: Failed to set IPv4 multicast
    ### [PREVIOUS LINE REPEATED 1 TIMES] ###
    May  4 08:19:55 nateserver wsdd[3851]: Failed to add multicast for WSDD: Address already in use
    May  4 08:19:55 nateserver wsdd[3851]: set_multicast: Failed to set IPv4 multicast
    ### [PREVIOUS LINE REPEATED 1 TIMES] ###
    May  4 08:19:55 nateserver winbindd[3855]: [2021/05/04 08:19:55.712654,  0] ../../source3/winbindd/winbindd_cache.c:3203(initialize_winbindd_cache)
    May  4 08:19:55 nateserver winbindd[3855]:   initialize_winbindd_cache: clearing cache and re-creating with version number 2
    May  4 08:19:55 nateserver winbindd[3855]: [2021/05/04 08:19:55.713023,  0] ../../lib/util/become_daemon.c:135(daemon_ready)
    May  4 08:19:55 nateserver winbindd[3855]:   daemon_ready: daemon 'winbindd' finished starting up and ready to serve connections
    May  4 08:19:55 nateserver root: Stopping docker_load
    May  4 08:19:55 nateserver emhttpd: shcmd (165): /etc/rc.d/rc.docker stop
    May  4 08:19:56 nateserver kernel: veth24961f6: renamed from eth0
    May  4 08:19:56 nateserver kernel: docker0: port 23(veth04b5f53) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Interface veth04b5f53.IPv6 no longer relevant for mDNS.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth04b5f53.IPv6 with address fe80::5c7a:bfff:fed8:8e4b.
    May  4 08:19:57 nateserver kernel: docker0: port 23(veth04b5f53) entered disabled state
    May  4 08:19:57 nateserver kernel: device veth04b5f53 left promiscuous mode
    May  4 08:19:57 nateserver kernel: docker0: port 23(veth04b5f53) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::5c7a:bfff:fed8:8e4b on veth04b5f53.
    May  4 08:19:57 nateserver kernel: veth512bbe1: renamed from eth0
    May  4 08:19:57 nateserver kernel: docker0: port 17(veth4dd0722) entered disabled state
    May  4 08:19:57 nateserver kernel: veth629bb14: renamed from eth0
    May  4 08:19:57 nateserver kernel: docker0: port 18(veth88eda13) entered disabled state
    May  4 08:19:57 nateserver kernel: veth15cfbff: renamed from eth0
    May  4 08:19:57 nateserver kernel: docker0: port 19(veth626b55c) entered disabled state
    May  4 08:19:57 nateserver kernel: veth9068688: renamed from eth0
    May  4 08:19:57 nateserver kernel: docker0: port 22(veth940e204) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Interface veth4dd0722.IPv6 no longer relevant for mDNS.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth4dd0722.IPv6 with address fe80::b0af:b3ff:fed3:585d.
    May  4 08:19:57 nateserver kernel: docker0: port 17(veth4dd0722) entered disabled state
    May  4 08:19:57 nateserver kernel: device veth4dd0722 left promiscuous mode
    May  4 08:19:57 nateserver kernel: docker0: port 17(veth4dd0722) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::b0af:b3ff:fed3:585d on veth4dd0722.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Interface veth626b55c.IPv6 no longer relevant for mDNS.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth626b55c.IPv6 with address fe80::f0c5:9fff:fef4:a070.
    May  4 08:19:57 nateserver kernel: docker0: port 19(veth626b55c) entered disabled state
    May  4 08:19:57 nateserver kernel: device veth626b55c left promiscuous mode
    May  4 08:19:57 nateserver kernel: docker0: port 19(veth626b55c) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::f0c5:9fff:fef4:a070 on veth626b55c.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Interface veth940e204.IPv6 no longer relevant for mDNS.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth940e204.IPv6 with address fe80::34b6:50ff:fe18:a1d4.
    May  4 08:19:57 nateserver kernel: docker0: port 22(veth940e204) entered disabled state
    May  4 08:19:57 nateserver kernel: device veth940e204 left promiscuous mode
    May  4 08:19:57 nateserver kernel: docker0: port 22(veth940e204) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::34b6:50ff:fe18:a1d4 on veth940e204.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Interface veth88eda13.IPv6 no longer relevant for mDNS.
    May  4 08:19:57 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth88eda13.IPv6 with address fe80::bc9f:c7ff:fe7a:f6f4.
    May  4 08:19:57 nateserver kernel: docker0: port 18(veth88eda13) entered disabled state
    May  4 08:19:57 nateserver kernel: device veth88eda13 left promiscuous mode
    May  4 08:19:57 nateserver kernel: docker0: port 18(veth88eda13) entered disabled state
    May  4 08:19:57 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::bc9f:c7ff:fe7a:f6f4 on veth88eda13.
    May  4 08:19:58 nateserver kernel: veth04276eb: renamed from eth0
    May  4 08:19:58 nateserver kernel: docker0: port 7(veth3a2b851) entered disabled state
    May  4 08:19:58 nateserver avahi-daemon[9066]: Interface veth3a2b851.IPv6 no longer relevant for mDNS.
    May  4 08:19:58 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth3a2b851.IPv6 with address fe80::3866:71ff:fe33:731.
    May  4 08:19:58 nateserver kernel: docker0: port 7(veth3a2b851) entered disabled state
    May  4 08:19:58 nateserver kernel: device veth3a2b851 left promiscuous mode
    May  4 08:19:58 nateserver kernel: docker0: port 7(veth3a2b851) entered disabled state
    May  4 08:19:58 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::3866:71ff:fe33:731 on veth3a2b851.
    May  4 08:20:00 nateserver kernel: veth62b5277: renamed from eth0
    May  4 08:20:00 nateserver kernel: veth235c620: renamed from eth0
    May  4 08:20:00 nateserver kernel: docker0: port 6(vethb3ec8f4) entered disabled state
    May  4 08:20:00 nateserver kernel: veth8478a91: renamed from eth0
    May  4 08:20:00 nateserver avahi-daemon[9066]: Interface vethb3ec8f4.IPv6 no longer relevant for mDNS.
    May  4 08:20:00 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vethb3ec8f4.IPv6 with address fe80::4837:23ff:febd:e3c8.
    May  4 08:20:00 nateserver kernel: docker0: port 6(vethb3ec8f4) entered disabled state
    May  4 08:20:00 nateserver kernel: device vethb3ec8f4 left promiscuous mode
    May  4 08:20:00 nateserver kernel: docker0: port 6(vethb3ec8f4) entered disabled state
    May  4 08:20:00 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::4837:23ff:febd:e3c8 on vethb3ec8f4.
    May  4 08:20:00 nateserver kernel: docker0: port 16(vetha078877) entered disabled state
    May  4 08:20:00 nateserver kernel: veth9bd4706: renamed from eth0
    May  4 08:20:00 nateserver avahi-daemon[9066]: Interface vetha078877.IPv6 no longer relevant for mDNS.
    May  4 08:20:00 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vetha078877.IPv6 with address fe80::b8e6:54ff:fe2a:3a4d.
    May  4 08:20:00 nateserver kernel: docker0: port 16(vetha078877) entered disabled state
    May  4 08:20:00 nateserver kernel: device vetha078877 left promiscuous mode
    May  4 08:20:00 nateserver kernel: docker0: port 16(vetha078877) entered disabled state
    May  4 08:20:00 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::b8e6:54ff:fe2a:3a4d on vetha078877.
    May  4 08:20:00 nateserver kernel: veth24b7eb1: renamed from eth0
    May  4 08:20:00 nateserver kernel: docker0: port 20(veth19440ff) entered disabled state
    May  4 08:20:00 nateserver kernel: veth2beb38e: renamed from eth0
    May  4 08:20:00 nateserver kernel: docker0: port 12(veth6655a77) entered disabled state
    May  4 08:20:01 nateserver kernel: docker0: port 11(vethb66a462) entered disabled state
    May  4 08:20:01 nateserver kernel: vethbdf3f05: renamed from eth0
    May  4 08:20:01 nateserver kernel: veth2300853: renamed from eth0
    May  4 08:20:01 nateserver kernel: docker0: port 14(veth4da7f84) entered disabled state
    May  4 08:20:01 nateserver kernel: docker0: port 21(veth42800d9) entered disabled state
    May  4 08:20:01 nateserver kernel: veth3807938: renamed from eth0
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface veth19440ff.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth19440ff.IPv6 with address fe80::603f:e3ff:fed5:88f0.
    May  4 08:20:01 nateserver kernel: docker0: port 20(veth19440ff) entered disabled state
    May  4 08:20:01 nateserver kernel: device veth19440ff left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 20(veth19440ff) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::603f:e3ff:fed5:88f0 on veth19440ff.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface veth6655a77.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth6655a77.IPv6 with address fe80::aca7:17ff:fe2e:8a35.
    May  4 08:20:01 nateserver kernel: docker0: port 12(veth6655a77) entered disabled state
    May  4 08:20:01 nateserver kernel: device veth6655a77 left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 12(veth6655a77) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::aca7:17ff:fe2e:8a35 on veth6655a77.
    May  4 08:20:01 nateserver kernel: veth01549aa: renamed from eth0
    May  4 08:20:01 nateserver kernel: vethdc2215b: renamed from eth0
    May  4 08:20:01 nateserver kernel: veth0e3f65a: renamed from eth0
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface veth4da7f84.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth4da7f84.IPv6 with address fe80::481f:f0ff:fe2d:8f2c.
    May  4 08:20:01 nateserver kernel: docker0: port 14(veth4da7f84) entered disabled state
    May  4 08:20:01 nateserver kernel: device veth4da7f84 left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 14(veth4da7f84) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::481f:f0ff:fe2d:8f2c on veth4da7f84.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface vethb66a462.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vethb66a462.IPv6 with address fe80::8cf2:e4ff:fef7:12e2.
    May  4 08:20:01 nateserver kernel: docker0: port 11(vethb66a462) entered disabled state
    May  4 08:20:01 nateserver kernel: device vethb66a462 left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 11(vethb66a462) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::8cf2:e4ff:fef7:12e2 on vethb66a462.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface veth42800d9.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth42800d9.IPv6 with address fe80::3cc0:76ff:fe5b:4da2.
    May  4 08:20:01 nateserver kernel: docker0: port 21(veth42800d9) entered disabled state
    May  4 08:20:01 nateserver kernel: device veth42800d9 left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 21(veth42800d9) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::3cc0:76ff:fe5b:4da2 on veth42800d9.
    May  4 08:20:01 nateserver kernel: vethdc998ad: renamed from eth0
    May  4 08:20:01 nateserver kernel: docker0: port 8(veth168c59c) entered disabled state
    May  4 08:20:01 nateserver kernel: veth6de98fe: renamed from eth0
    May  4 08:20:01 nateserver kernel: veth18b0952: renamed from eth0
    May  4 08:20:01 nateserver kernel: docker0: port 9(vethea7cb35) entered disabled state
    May  4 08:20:01 nateserver kernel: docker0: port 3(veth0767946) entered disabled state
    May  4 08:20:01 nateserver kernel: veth6697af3: renamed from eth0
    May  4 08:20:01 nateserver kernel: vethacedea6: renamed from eth0
    May  4 08:20:01 nateserver kernel: docker0: port 10(veth7627f42) entered disabled state
    May  4 08:20:01 nateserver kernel: docker0: port 13(vethdb63b55) entered disabled state
    May  4 08:20:01 nateserver kernel: veth9f94df7: renamed from eth0
    May  4 08:20:01 nateserver kernel: docker0: port 2(vetha75e365) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Interface veth168c59c.IPv6 no longer relevant for mDNS.
    May  4 08:20:01 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth168c59c.IPv6 with address fe80::9d:4dff:fe56:a32c.
    May  4 08:20:01 nateserver kernel: docker0: port 8(veth168c59c) entered disabled state
    May  4 08:20:01 nateserver kernel: device veth168c59c left promiscuous mode
    May  4 08:20:01 nateserver kernel: docker0: port 8(veth168c59c) entered disabled state
    May  4 08:20:01 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::9d:4dff:fe56:a32c on veth168c59c.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Interface vethea7cb35.IPv6 no longer relevant for mDNS.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vethea7cb35.IPv6 with address fe80::18c7:15ff:fe22:7832.
    May  4 08:20:02 nateserver kernel: docker0: port 9(vethea7cb35) entered disabled state
    May  4 08:20:02 nateserver kernel: device vethea7cb35 left promiscuous mode
    May  4 08:20:02 nateserver kernel: docker0: port 9(vethea7cb35) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::18c7:15ff:fe22:7832 on vethea7cb35.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Interface veth0767946.IPv6 no longer relevant for mDNS.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth0767946.IPv6 with address fe80::3885:8cff:fec6:e92d.
    May  4 08:20:02 nateserver kernel: docker0: port 3(veth0767946) entered disabled state
    May  4 08:20:02 nateserver kernel: device veth0767946 left promiscuous mode
    May  4 08:20:02 nateserver kernel: docker0: port 3(veth0767946) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::3885:8cff:fec6:e92d on veth0767946.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Interface vethdb63b55.IPv6 no longer relevant for mDNS.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vethdb63b55.IPv6 with address fe80::fc8e:8dff:fe82:e15c.
    May  4 08:20:02 nateserver kernel: docker0: port 13(vethdb63b55) entered disabled state
    May  4 08:20:02 nateserver kernel: device vethdb63b55 left promiscuous mode
    May  4 08:20:02 nateserver kernel: docker0: port 13(vethdb63b55) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::fc8e:8dff:fe82:e15c on vethdb63b55.
    May  4 08:20:02 nateserver kernel: docker0: port 10(veth7627f42) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Interface veth7627f42.IPv6 no longer relevant for mDNS.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth7627f42.IPv6 with address fe80::c898:afff:fe9e:412e.
    May  4 08:20:02 nateserver kernel: device veth7627f42 left promiscuous mode
    May  4 08:20:02 nateserver kernel: docker0: port 10(veth7627f42) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::c898:afff:fe9e:412e on veth7627f42.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Interface vetha75e365.IPv6 no longer relevant for mDNS.
    May  4 08:20:02 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vetha75e365.IPv6 with address fe80::8454:52ff:feff:628.
    May  4 08:20:02 nateserver kernel: docker0: port 2(vetha75e365) entered disabled state
    May  4 08:20:02 nateserver kernel: device vetha75e365 left promiscuous mode
    May  4 08:20:02 nateserver kernel: docker0: port 2(vetha75e365) entered disabled state
    May  4 08:20:02 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::8454:52ff:feff:628 on vetha75e365.
    May  4 08:20:05 nateserver kernel: veth761f5dc: renamed from eth0
    May  4 08:20:06 nateserver kernel: veth4d76f90: renamed from eth0
    May  4 08:20:06 nateserver kernel: docker0: port 15(veth6cfedc5) entered disabled state
    May  4 08:20:06 nateserver kernel: veth3b89158: renamed from eth0
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface veth6cfedc5.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth6cfedc5.IPv6 with address fe80::9862:68ff:fe8c:8c54.
    May  4 08:20:06 nateserver kernel: docker0: port 15(veth6cfedc5) entered disabled state
    May  4 08:20:06 nateserver kernel: device veth6cfedc5 left promiscuous mode
    May  4 08:20:06 nateserver kernel: docker0: port 15(veth6cfedc5) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::9862:68ff:fe8c:8c54 on veth6cfedc5.
    May  4 08:20:06 nateserver kernel: veth6ab0901: renamed from eth0
    May  4 08:20:06 nateserver kernel: docker0: port 1(vethb1ba683) entered disabled state
    May  4 08:20:06 nateserver kernel: veth77351b9: renamed from eth0
    May  4 08:20:06 nateserver kernel: docker0: port 5(veth404ff80) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface vethb1ba683.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vethb1ba683.IPv6 with address fe80::6c12:2fff:fe8f:da75.
    May  4 08:20:06 nateserver kernel: docker0: port 1(vethb1ba683) entered disabled state
    May  4 08:20:06 nateserver kernel: device vethb1ba683 left promiscuous mode
    May  4 08:20:06 nateserver kernel: docker0: port 1(vethb1ba683) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::6c12:2fff:fe8f:da75 on vethb1ba683.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface veth404ff80.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth404ff80.IPv6 with address fe80::5ce9:49ff:fe0a:a828.
    May  4 08:20:06 nateserver kernel: docker0: port 5(veth404ff80) entered disabled state
    May  4 08:20:06 nateserver kernel: device veth404ff80 left promiscuous mode
    May  4 08:20:06 nateserver kernel: docker0: port 5(veth404ff80) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::5ce9:49ff:fe0a:a828 on veth404ff80.
    May  4 08:20:06 nateserver kernel: veth1713b29: renamed from eth0
    May  4 08:20:06 nateserver kernel: docker0: port 4(veth43d7fc1) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface veth43d7fc1.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface veth43d7fc1.IPv6 with address fe80::94d7:e4ff:feb1:6171.
    May  4 08:20:06 nateserver kernel: docker0: port 4(veth43d7fc1) entered disabled state
    May  4 08:20:06 nateserver kernel: device veth43d7fc1 left promiscuous mode
    May  4 08:20:06 nateserver kernel: docker0: port 4(veth43d7fc1) entered disabled state
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::94d7:e4ff:feb1:6171 on veth43d7fc1.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for 192.168.1.6 on shim-br0.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.IPv4 with address 192.168.1.6.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface shim-br0.IPv4 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for 2600:8801:dd00:880:18ec:aaff:fe0c:fd58 on shim-br0.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.IPv6 with address 2600:8801:dd00:880:18ec:aaff:fe0c:fd58.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Joining mDNS multicast group on interface shim-br0.IPv6 with address fe80::18ec:aaff:fe0c:fd58.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Registering new address record for fe80::18ec:aaff:fe0c:fd58 on shim-br0.*.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::18ec:aaff:fe0c:fd58 on shim-br0.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.IPv6 with address fe80::18ec:aaff:fe0c:fd58.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface shim-br0.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver kernel: device br0 left promiscuous mode
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for 192.168.3.61 on shim-br0.3.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.3.IPv4 with address 192.168.3.61.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface shim-br0.3.IPv4 no longer relevant for mDNS.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for 2600:8801:dd00:881:845a:b3ff:feef:38ae on shim-br0.3.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.3.IPv6 with address 2600:8801:dd00:881:845a:b3ff:feef:38ae.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Joining mDNS multicast group on interface shim-br0.3.IPv6 with address fe80::845a:b3ff:feef:38ae.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Registering new address record for fe80::845a:b3ff:feef:38ae on shim-br0.3.*.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::845a:b3ff:feef:38ae on shim-br0.3.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface shim-br0.3.IPv6 with address fe80::845a:b3ff:feef:38ae.
    May  4 08:20:06 nateserver avahi-daemon[9066]: Interface shim-br0.3.IPv6 no longer relevant for mDNS.
    May  4 08:20:06 nateserver kernel: device br0.3 left promiscuous mode
    May  4 08:20:06 nateserver root: stopping dockerd ...
    May  4 08:20:07 nateserver root: waiting for docker to die ...
    May  4 08:20:08 nateserver avahi-daemon[9066]: Interface docker0.IPv6 no longer relevant for mDNS.
    May  4 08:20:08 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface docker0.IPv6 with address fe80::42:9fff:fe90:eac1.
    May  4 08:20:08 nateserver avahi-daemon[9066]: Interface docker0.IPv4 no longer relevant for mDNS.
    May  4 08:20:08 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface docker0.IPv4 with address 172.17.0.1.
    May  4 08:20:08 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::42:9fff:fe90:eac1 on docker0.
    May  4 08:20:08 nateserver avahi-daemon[9066]: Withdrawing address record for 172.17.0.1 on docker0.
    May  4 08:20:08 nateserver emhttpd: shcmd (166): umount /var/lib/docker
    May  4 08:20:10 nateserver nmbd[3844]: [2021/05/04 08:20:10.702312,  0] ../../source3/libsmb/nmblib.c:922(send_udp)
    May  4 08:20:10 nateserver nmbd[3844]:   Packet send failed to 172.17.255.255(138) ERRNO=Network is unreachable
    May  4 08:20:10 nateserver nmbd[3844]: [2021/05/04 08:20:10.702366,  0] ../../source3/libsmb/nmblib.c:922(send_udp)
    May  4 08:20:10 nateserver nmbd[3844]:   Packet send failed to 172.17.255.255(137) ERRNO=Network is unreachable
    May  4 08:20:10 nateserver nmbd[3844]: [2021/05/04 08:20:10.702376,  0] ../../source3/nmbd/nmbd_packets.c:179(send_netbios_packet)
    May  4 08:20:10 nateserver nmbd[3844]:   send_netbios_packet: send_packet() to IP 172.17.255.255 port 137 failed
    May  4 08:20:10 nateserver nmbd[3844]: [2021/05/04 08:20:10.702388,  0] ../../source3/nmbd/nmbd_nameregister.c:581(register_name)
    May  4 08:20:10 nateserver nmbd[3844]:   register_name: Failed to send packet trying to register name #001#002__MSBROWSE__#002<01>
    May  4 08:20:18 nateserver nmbd[3844]: [2021/05/04 08:20:18.710436,  0] ../../source3/nmbd/nmbd_become_lmb.c:397(become_local_master_stage2)
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   Samba name server NATESERVER is now a local master browser for workgroup WORKGROUP on subnet 192.168.1.6
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:20:18 nateserver nmbd[3844]: [2021/05/04 08:20:18.710496,  0] ../../source3/nmbd/nmbd_become_lmb.c:397(become_local_master_stage2)
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   Samba name server NATESERVER is now a local master browser for workgroup WORKGROUP on subnet 192.168.3.61
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:20:18 nateserver nmbd[3844]: [2021/05/04 08:20:18.710530,  0] ../../source3/nmbd/nmbd_become_lmb.c:397(become_local_master_stage2)
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   Samba name server NATESERVER is now a local master browser for workgroup WORKGROUP on subnet 192.168.122.1
    May  4 08:20:18 nateserver nmbd[3844]:   
    May  4 08:20:18 nateserver nmbd[3844]:   *****
    May  4 08:21:08 nateserver nmbd[3844]: [2021/05/04 08:21:08.760377,  0] ../../source3/libsmb/nmblib.c:922(send_udp)
    May  4 08:21:08 nateserver nmbd[3844]:   Packet send failed to 172.17.255.255(138) ERRNO=Network is unreachable
    May  4 08:23:42 nateserver kernel: usb 7-2: new low-speed USB device number 4 using xhci_hcd
    May  4 08:23:42 nateserver kernel: input: NOVATEK USB Keyboard as /devices/pci0000:00/0000:00:08.1/0000:0f:00.3/usb7/7-2/7-2:1.0/0003:0461:0010.000F/input/input16
    May  4 08:23:42 nateserver kernel: hid-generic 0003:0461:0010.000F: input,hidraw0: USB HID v1.10 Keyboard [NOVATEK USB Keyboard] on usb-0000:0f:00.3-2/input0
    May  4 08:23:42 nateserver kernel: input: NOVATEK USB Keyboard System Control as /devices/pci0000:00/0000:00:08.1/0000:0f:00.3/usb7/7-2/7-2:1.1/0003:0461:0010.0010/input/input17
    May  4 08:23:42 nateserver kernel: input: NOVATEK USB Keyboard Consumer Control as /devices/pci0000:00/0000:00:08.1/0000:0f:00.3/usb7/7-2/7-2:1.1/0003:0461:0010.0010/input/input18
    May  4 08:23:42 nateserver kernel: hid-generic 0003:0461:0010.0010: input,hidraw1: USB HID v1.10 Device [NOVATEK USB Keyboard] on usb-0000:0f:00.3-2/input1
    May  4 08:23:58 nateserver login: pam_unix(login:session): session opened for user root(uid=0) by LOGIN(uid=0)
    May  4 08:23:58 nateserver login: ROOT LOGIN ON tty1
    May  4 08:27:55 nateserver avahi-daemon[9066]: Interface vnet0.IPv6 no longer relevant for mDNS.
    May  4 08:27:55 nateserver avahi-daemon[9066]: Leaving mDNS multicast group on interface vnet0.IPv6 with address fe80::fc54:ff:fee9:c18a.
    May  4 08:27:55 nateserver kernel: br0: port 2(vnet0) entered disabled state
    May  4 08:27:55 nateserver kernel: device vnet0 left promiscuous mode
    May  4 08:27:55 nateserver kernel: br0: port 2(vnet0) entered disabled state
    May  4 08:27:55 nateserver avahi-daemon[9066]: Withdrawing address record for fe80::fc54:ff:fee9:c18a on vnet0.
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: version 1.1
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: performing pre-reset
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: performing reset
    May  4 08:27:57 nateserver kernel: ATOM BIOS: 113-EXT40650-001
    May  4 08:27:57 nateserver kernel: vendor-reset-drm: atomfirmware: bios_scratch_reg_offset initialized to 4c
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: bus reset disabled? yes
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: SMU response reg: 1, sol reg: 19fa8b38, mp1 intr enabled? yes, bl ready? yes
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: Clearing scratch regs 6 and 7
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: begin psp mode 1 reset
    May  4 08:27:57 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: mode1 reset succeeded
    May  4 08:27:59 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: PSP mode1 reset successful
    May  4 08:27:59 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: performing post-reset
    May  4 08:27:59 nateserver kernel: vfio-pci 0000:0d:00.0: AMD_NAVI10: reset result = 0
    

     

    nateserver-diagnostics-20210504-0824.zip

  10. 9 hours ago, giganode said:

     

    I am running a custom build with nvidia support and the vendor-reset.

    Please do a new build with the kernel-helper but integrate nvidia support into it. Remove the nvidia plugin if you readded it.

     

     

     

    12 hours ago, ich777 said:

    You can try it but please keep in mind that this could lead to the same problem. Please report back.

     

    Both the Nvidia drivers and the vender-reset are working; there were no issues booting up. Thank you so much for your help and hard work, I really appreciate it! If you need someone to test future changes or updates don't hesitate to reach out.

    • Like 2
  11. 43 minutes ago, ich777 said:

    Have you built the Images with Nvidia or do you use the Nvidia-Driver Plugin (but I think the filesize tells me that you built it without the Nvidia drivers)?

     

    Please try to uninstall the Nvidia-Driver Plugin for troubleshooting reasons and then try to boot with the custom images.

     

    Have you also installed any other modifications like a custom Kernel (bzimage) or something?

     

    13 minutes ago, ndetar said:

    So I just booted up without the Nvidia Driver Plugin and it started up with no issues.

     

    I also just tested shutting down and starting up a VM and the gnif/vender-reset seems to be working correctly! The only thing left would be to figure out the Nvidia drivers issue. I have a second Nvidia GPU I use for transcoding. Should I try building the kernel with the Nvidia drivers built in rather than the plugin?

    • Like 1
  12. 28 minutes ago, ich777 said:

    Have you built the Images with Nvidia or do you use the Nvidia-Driver Plugin (but I think the filesize tells me that you built it without the Nvidia drivers)?

     

    Please try to uninstall the Nvidia-Driver Plugin for troubleshooting reasons and then try to boot with the custom images.

     

    Have you also installed any other modifications like a custom Kernel (bzimage) or something?

    So I just booted up without the Nvidia Driver Plugin and it started up with no issues.

    • Like 1
  13. 3 minutes ago, ich777 said:

    Have you built the Images with Nvidia or do you use the Nvidia-Driver Plugin (but I think the filesize tells me that you built it without the Nvidia drivers)?

     

    Please try to uninstall the Nvidia-Driver Plugin for troubleshooting reasons and then try to boot with the custom images.

     

    Have you also installed any other modifications like a custom Kernel (bzimage) or something?

    I use the Nvidia driver plugin and no other modifications. I'll try it again without the Nvidia driver plugin.

    • Thanks 1
  14. 15 hours ago, ich777 said:

    Have you something appended to your syslinux.cfg?

    Eventually @giganode can help you out.

     

    What you can try in the meantime is delete the container entirely and the output folder that lives in the kernel directory in your appdata folder and redownload the Unraid-Kernel-Helper from the CA App the new Variable that @giganode mentioned should be already in there.

     

    Then select to build it with the gnif/vendor-reset patch and also click on 'Show more settings...' and at gnif/vendor-reset Branch enter 'feature/audio_reset' and also set to option Save to Log to 'true'.

     

    If that all doesn't work please post the Log file that it creates in the Kernel directory.

     

    13 hours ago, giganode said:

    I can share my settings, but I feel like something went wrong with the build.. But I don't know for a fact :) 

     

    kernel /bzimage
    append video=vesafb:off,efifb:off vfio_iommu_type1.allow_unsafe_interrupts=1 isolcpus=4-15,20-31 pcie_acs_override=downstream,multifunction initrd=/bzroot

    So I rebuilt the image and tried it again, same problem as before. This time I tried to ssh into the server and was able to grab the syslog, the diagnostics command never completed. I've also attached the log from the build.

    syslog.txt 2021-01-14_09.16.44.log

  15. 9 hours ago, ich777 said:

    Have you something appended to your syslinux.cfg?

    Eventually @giganode can help you out.

     

    What you can try in the meantime is delete the container entirely and the output folder that lives in the kernel directory in your appdata folder and redownload the Unraid-Kernel-Helper from the CA App the new Variable that @giganode mentioned should be already in there.

     

    Then select to build it with the gnif/vendor-reset patch and also click on 'Show more settings...' and at gnif/vendor-reset Branch enter 'feature/audio_reset' and also set to option Save to Log to 'true'.

     

    If that all doesn't work please post the Log file that it creates in the Kernel directory.

     

    8 hours ago, giganode said:

    I can share my settings, but I feel like something went wrong with the build.. But I don't know for a fact :) 

     

    kernel /bzimage
    append video=vesafb:off,efifb:off vfio_iommu_type1.allow_unsafe_interrupts=1 isolcpus=4-15,20-31 pcie_acs_override=downstream,multifunction initrd=/bzroot

     

    Here is my syslinux.cfg I've only marked some cores for isolation.

    kernel /bzimage
    append isolcpus=3-7,11-15 initrd=/bzroot

     

    I agree it is likely the build, I'll be rebuilding it later today and post the log. Thanks for the input.

  16. 2 hours ago, giganode said:

    The vendor-reset got an update for navi users. I tested it on my system.

    I have no broken audio anymore after resets. For me this is a real breakthrough!

    I don't need the old navi patch anymore now.

     

    I can now boot between Windows 10 20H2, macOS Big Sur 11.1 and Ubuntu 20.10.

     

    For all navi user who wanna try it out:

     

    1. Force update the docker
    2. Edit the docker and add a variable like this:image.thumb.png.39429a29f1c092625e8a643ca9823ebd.png
    3. Try and hopefully enjoy!

    Keep in mind, this only fixes the specific audio issue for navi users.

     

    Please report your expierences here. :) 

     

    Special Thanks to @ich777 for that fast edit.

     

     

     

    I have a 5600XT. This is my first time trying the gnif/vender-reset (in the past I used the navi reset patch), but when I tried it Unraid failed to startup, it just stopped at this point. I am pretty sure the compile completed correctly, got the usual all done message. More than willing to do more testing/troubleshooting.

     

    20210113_145700.thumb.jpg.df9d31383f3b67e1108d8e8ac6dad3e3.jpg

  17. On 1/6/2021 at 5:51 PM, jonathanm said:

     

    On 1/7/2021 at 7:59 AM, ghost82 said:

    I think you can move the primary gpu to slot 1, and the 5600xt to any other slot.

    If I look at my system, the system is using slot 1 gpu to boot.

    In any case attach the 5600xt to vfio.

    Thanks for the help and recommendations! I swapped around the graphics cards, but was still having issues so I did a clean windows install and that seems to have fixed all of my issues. I really appreciate the help!

    • Like 1
  18. 2 minutes ago, ghost82 said:

    I think you can move the primary gpu to slot 1, and the 5600xt to any other slot.

    If I look at my system, the os is using slot 1 gpu to boot.

    In any case attach the 5600xt to vfio.

    OK thanks. I'll look into switching them around. Hopefully I can, the 5600XT is a large card. I won't be able to do that for a little while but once I do I'll update.

  19. 1 minute ago, ghost82 said:

    I think so, because from what you write you can see unraid is binding to the 5600xt, if you want to use igpu set it in the bios as the primary gpu and use vfio for attaching the 5600xt.

    However you will not get rid of the reset bug if you will use the 5600xt for your vm(s), unless you use some "tricks" to reset the card after shutting down/rebooting the vm.

    Well since I switched to the 3700X I no longer have an igpu. I have a secondary NVIDIA gpu, is there any way to get unraid to bind to that? I was unable to find a setting in the bios to switch primary gpus.

  20. 8 hours ago, ghost82 said:

    They reset the gpu (commands unhook unraid from the gpu, so it can be properly passed through to the vm), search for amd reset bug.

    I am aware of the reset bug, had to deal with it before switching hardware, however I didn't have to use those commands before. Is this due to Unraid binding to the 5600XT where as before Unraid would bind to the intel iGPU?

  21. So, strange stuff has happened. For some reason my server isn't logging anything, no idea why I'll have to figure that out later.

     

    I decided to create a new unraid usb. Kinda start from scratch. I started a vm with the 5600XT passed through and the system log was filled up with:

    unraid kernel: vfio-pci 0000:0c:00.0: BAR 0: can't reserve [mem 0xd0000000-0xdfffffff 64bit pref]

    Which lead me to this thread: https://forums.unraid.net/topic/77241-unable-to-passthrough-primary-gpu/

    Which recommended running the following commands:

    echo 0 > /sys/class/vtconsole/vtcon0/bind
    echo 0 > /sys/class/vtconsole/vtcon1/bind
    echo efi-framebuffer.0 > /sys/bus/platform/drivers/efi-framebuffer/unbind

    After running those commands on a fresh reboot the VM started up with no problem. So I switched back to my original usb drive and ran the same commands before starting the vm and it started right up.

     

    I honestly don't know what those commands do so if there is some other more proper fix/solution or an explanation as to what those commands do I'd be very appreciative.

  22. 19 minutes ago, jonathanm said:

    My hunch was based on the video corruption, so my next guess would involve the video BIOS you dumped. What happens if you start the VM without the vbios specified?

     

    15 minutes ago, ndetar said:

    The same thing. I think the video corruption might actually be incorrectly displayed console text, but i'm not completly sure. I recently noticed that if I hit enter or type the pixelation shifts.

    Additionally the last time I rebooted the pixelation started shift up the screen like it was just incorrectly rendering the console reboot text.