NeoDude

Members
  • Posts

    298
  • Joined

  • Last visited

Everything posted by NeoDude

  1. No more errors with my new CPUs
  2. Ok, I think I'm missing something here. I have had this setup for a while, as per the Spaceinvaderone video, with all of my appdata folders as datasets, using the script to create snapshots and replicate to a ZFS disk in the array. Today is the first time I've had to try and rollback/restore and I'm at a loss. So, my calibre-web install appears to have reverted to a new install, and I can't even login. No problem I thought, I'll rollback to a snapshot from last week, when I knew it was working. Doing this results in an empty appdata folder. Weird I thought, but no problem, I have these replicated. And so my first issue, there is no documentation anywhere on how I get my replicated snapshots from Disc1 back to the Cache. Do I even need to do this? Can I not just restore the appdata folder from Disk1? Any ideas why rolling back the snapshots on Cache are resulting in empty folders? So confused.
  3. What's the general consensus or best practice with whether or not to have recordings going via SSD Cache Pool or Straight to array? Is SSD wear with constant writes still a consideration with modern drives?
  4. Thanks for the reply, nothing extra in the Enhanced Log. I'm going to upgrade to a pair of E5-2697s in the next month so will see what happens then.
  5. Perhaps, but it's been happening consistently up until today. I'll keep an eye on it
  6. Weird. When I rebooted into safe mode, everything unmounted fine. Both rebooting into safe mode, and then rebooting normally. No issues.
  7. This seems to be an ongoing issue, whenever I try to stop the Array, I get this repeating in the log... Aug 7 12:49:52 Gandalf root: cannot unmount '/mnt/cache/system': pool or dataset is busy Aug 7 12:49:52 Gandalf emhttpd: shcmd (336): exit status: 1 Aug 7 12:49:52 Gandalf emhttpd: Retry unmounting disk share(s)... This also results in a Parity Check after every reboot. I've seen others with the issue solve it by updating to the latest version, but I'm already on the latest version. Any ideas? Diags Attached. gandalf-diagnostics-20230807-1248.zip
  8. I'm getting this error repeated in my syslog over and over every minute or 2, nothing else seems to be affected and the server is rock solid stable. I have carried out an overnight memory test also, without issue. Any ideas... Jul 15 09:11:51 Gandalf kernel: mce: [Hardware Error]: Machine check events logged Jul 15 09:11:51 Gandalf mcelog: Running trigger `bus-error-trigger' (reporter: bus) Jul 15 09:11:51 Gandalf mcelog: CPU 8 on socket 1 received Bus and Interconnect Errors in Other-transaction Jul 15 09:11:51 Gandalf mcelog: Location: CPU 8 on socket 1 Diagnostics Attached. Thanks in advance for any insights gandalf-diagnostics-20230715-1026.zip
  9. Just a minor one. I have 2 GPUs in my system, a Quadro P600, and a GeForce GTX 1050. I have the 1050 successfully working with Frigate. 'Nvidia-SMI' shows that the 1050 is being used by ffmpeg, but the GUI in Frigate has the P600 listed... Any ideas?
  10. Think I found the issue. There was a missing underscore in the "NVIDIA_VISIBLE_DEVICES" key. Not sure if this is default on the container or if it's something I've accidently done, probably the latter
  11. I've deleted the unrequired VFIO Bindings. These weren't checked in the GUI so I don't know why they were in there. I have also disabled Privileged mode. (This was a recent thing to see if it made a difference). After a reboot, Plex is now using the correct GPU, but TDARR is not. Here's the Docker Run for TDARR... docker run -d --name='tdarr' --net='br0.50' --ip='172.16.50.250' --cpuset-cpus='2,3,4,5,18,19,20,21' -e TZ="Europe/London" -e HOST_OS="Unraid" -e HOST_HOSTNAME="Gandalf" -e HOST_CONTAINERNAME="tdarr" -e 'serverIP'='172.16.50.250' -e 'TCP_PORT_8266'='8266' -e 'TCP_PORT_8265'='8265' -e 'PUID'='99' -e 'PGID'='100' -e 'internalNode'='true' -e 'NVIDIA_VISIBLE DEVICES'='GPU-04dd732e-60ad-a070-80b2-a0c4f284a9c1' -e 'NVIDIA_DRIVER_CAPABILITIES'='all' -e 'nodeIP'='0.0.0.0' -e 'nodeID'='Gandalf' -e 'TCP_PORT_8264'='8264' -l net.unraid.docker.managed=dockerman -l net.unraid.docker.webui='http://[IP]:[PORT:8265]' -l net.unraid.docker.icon='https://raw.githubusercontent.com/selfhosters/unRAID-CA-templates/master/templates/img/tdarr.png' -v '/mnt/user/appdata/tdarr/server':'/app/server':'rw' -v '/mnt/user/appdata/tdarr/configs':'/app/configs':'rw' -v '/mnt/user/appdata/tdarr/logs':'/app/logs':'rw' -v '/mnt/user0/media/':'/media':'rw' -v '/mnt/cache/appdata/tdarr/temp/':'/temp':'rw' --runtime=nvidia 'haveagitgat/tdarr_acc:dev' 2f5017a5896ff9f586419bf25d1a736256d750b6e2e8c97a2fb2f96b22597c2a
  12. I apparently posted this in the wrong place (I wasn't going to assume that it was an issue with this Plugin). So here it is again, as requested... Hey folks, I have a strange issue. I have 2 GPUs installed in my system, a Quadro P600, and a GTX 1050. These are both recognised by the Nvidia Driver plugin. I have the correct ID setup in both the TDARR and Plex docker containers for the P600, but they both still insist on using the 1050. Any Ideas? gandalf-diagnostics-20230505-2131.zip
  13. Hey folks, I have a strange issue. I have 2 GPUs installed in my system, a Quadro P600, and a GTX 1050. These are both recognised by the Nvidia Driver plugin. I have the correct ID setup in both the TDARR and Plex docker containers for the P600, but they both still insist on using the 1050. Any Ideas?
  14. Kinda defeats the point of my current setup, but thanks for your efforts. I'll just live with it.
  15. Yes. Every other device has full connectivity.
  16. These diags were during an outage... gandalf-diagnostics-20220907-0810.zip
  17. It happens randomly a few times every hour. I'll open terminal and try to ping a random website, half the time it doesn't work, the other half the problem has fixed itself in the time it takes me to open terminal and type.
  18. Had this issue for a few weeks now. I seem to get random Network disconnects for no apparent reason. I get the Exclamation Mark in the header and when I click it an error message, either "Relay Returned a 500 error" or something about not being able to resolve https://mothership.unraid.net. I also cannot access community apps or update containers while this error shows. It is completely random, and it usually fixes itself after a few minutes. Any ideas where to start looking for the cause? Diagnostics attached. gandalf-diagnostics-20220902-1409.zip
  19. Installed this today. Works perfectly using VNC, but I get no display output when I attempt to pass through my RX480. The GPU is confirmed good with my other Ubuntu VM. Any ideas? *EDIT* Further Info. I get the following errors when starting my VM... Sep 3 15:13:47 Gandalf kernel: DMAR: VT-d detected Invalidation Completion Error: SID 0 Sep 3 15:13:47 Gandalf kernel: DMAR: QI HEAD: IOTLB Invalidation qw0 = 0x200f2, qw1 = 0xdc0000e Sep 3 15:13:47 Gandalf kernel: DMAR: QI PRIOR: Invalidation Wait qw0 = 0x200000025, qw1 = 0x1001114dc Sep 3 15:13:47 Gandalf kernel: DMAR: Invalidation Completion Error (ICE) cleared But I also get them on my Ubuntu VM and it subsequently boots up fine. I have also tried to dump the vBios, which fails and gives this error... Sep 3 15:18:52 Gandalf kernel: vfio-pci 0000:04:00.0: Invalid PCI ROM header signature: expecting 0xaa55, got 0xffff XML... <?xml version='1.0' encoding='UTF-8'?> <domain type='kvm' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'> <name>BigSur</name> <uuid>*****</uuid> <description>MacOS Big Sur</description> <metadata> <vmtemplate xmlns="unraid" name="Windows 10" icon="Apple_vintage_trans.png" os="osx"/> </metadata> <memory unit='KiB'>16777216</memory> <currentMemory unit='KiB'>16777216</currentMemory> <memoryBacking> <nosharepages/> </memoryBacking> <vcpu placement='static'>8</vcpu> <cputune> <vcpupin vcpu='0' cpuset='12'/> <vcpupin vcpu='1' cpuset='28'/> <vcpupin vcpu='2' cpuset='13'/> <vcpupin vcpu='3' cpuset='29'/> <vcpupin vcpu='4' cpuset='14'/> <vcpupin vcpu='5' cpuset='30'/> <vcpupin vcpu='6' cpuset='15'/> <vcpupin vcpu='7' cpuset='31'/> </cputune> <os> <type arch='x86_64' machine='pc-q35-6.2'>hvm</type> <loader readonly='yes' type='pflash'>/mnt/user/system/custom_ovmf/Macinabox_CODE-pure-efi.fd</loader> <nvram>/etc/libvirt/qemu/nvram/*****_VARS-pure-efi.fd</nvram> </os> <features> <acpi/> <apic/> </features> <cpu mode='host-passthrough' check='none' migratable='on'> <topology sockets='1' dies='1' cores='4' threads='2'/> <cache mode='passthrough'/> </cpu> <clock offset='utc'> <timer name='rtc' tickpolicy='catchup'/> <timer name='pit' tickpolicy='delay'/> <timer name='hpet' present='no'/> </clock> <on_poweroff>destroy</on_poweroff> <on_reboot>restart</on_reboot> <on_crash>restart</on_crash> <devices> <emulator>/usr/local/sbin/qemu</emulator> <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='writeback'/> <source file='/mnt/user/vDisks/Macinabox BigSur/macos_disk.img'/> <target dev='hdc' bus='sata'/> <boot order='1'/> <address type='drive' controller='0' bus='0' target='0' unit='2'/> </disk> <controller type='pci' index='0' model='pcie-root'/> <controller type='pci' index='1' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='1' port='0x9'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> </controller> <controller type='pci' index='2' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='2' port='0xa'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> </controller> <controller type='pci' index='3' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='3' port='0xc'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x4'/> </controller> <controller type='pci' index='4' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='4' port='0x13'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x3'/> </controller> <controller type='pci' index='5' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='5' port='0x8'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='6' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='6' port='0xb'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x3'/> </controller> <controller type='pci' index='7' model='pcie-to-pci-bridge'> <model name='pcie-pci-bridge'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </controller> <controller type='virtio-serial' index='0'> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </controller> <controller type='sata' index='0'> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/> </controller> <controller type='usb' index='0' model='ich9-ehci1'> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x7'/> </controller> <controller type='usb' index='0' model='ich9-uhci1'> <master startport='0'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0' multifunction='on'/> </controller> <controller type='usb' index='0' model='ich9-uhci2'> <master startport='2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x1'/> </controller> <controller type='usb' index='0' model='ich9-uhci3'> <master startport='4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x2'/> </controller> <interface type='bridge'> <mac address='52:54:00:e0:a3:b7'/> <source bridge='br0'/> <model type='virtio'/> <address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/> </interface> <serial type='pty'> <target type='isa-serial' port='0'> <model name='isa-serial'/> </target> </serial> <console type='pty'> <target type='serial' port='0'/> </console> <channel type='unix'> <target type='virtio' name='org.qemu.guest_agent.0'/> <address type='virtio-serial' controller='0' bus='0' port='1'/> </channel> <input type='tablet' bus='usb'> <address type='usb' bus='0' port='1'/> </input> <input type='mouse' bus='ps2'/> <input type='keyboard' bus='ps2'/> <audio id='1' type='none'/> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x04' slot='0x00' function='0x0'/> </source> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x04' slot='0x00' function='0x1'/> </source> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x84' slot='0x00' function='0x0'/> </source> <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0'/> </hostdev> <memballoon model='none'/> </devices> <qemu:commandline> <qemu:arg value='-usb'/> <qemu:arg value='-device'/> <qemu:arg value='usb-kbd,bus=usb-bus.0'/> <qemu:arg value='-device'/> <qemu:arg value='************************'/> <qemu:arg value='-smbios'/> <qemu:arg value='type=2'/> <qemu:arg value='-cpu'/> <qemu:arg value='Penryn,kvm=on,vendor=GenuineIntel,+kvm_pv_unhalt,+kvm_pv_eoi,+hypervisor,+invtsc,+pcid,+ssse3,+sse4.2,+popcnt,+avx,+avx2,+aes,+fma,+fma4,+bmi1,+bmi2,+xsave,+xsaveopt,+rdrand,check'/> </qemu:commandline> </domain>
  20. I'm assuming that this is the reason that none of my websites are accessible today. Any ideas how to fix it? *EDIT* Disregard. Fixed it. Appears that the new version of Nginx requires Conf files to have the .conf extension or it won't load them.
  21. Ok, I fixed it. The issue was due to the fact I had a Static IP mapping entry for the Container in PfSense. The MAC address must've changed (possibly when I changed from MACVLAN to IPVLAN?). Deleted it and all is good with the world. Thanks for steering me in the right direction