richland007

Members
  • Posts

    26
  • Joined

  • Last visited

Posts posted by richland007

  1. @SimonF So go i just add this passage to the xml ?

    this is what @ghost82 wrote above for a Q35 machine:

     <hostdev mode='subsystem' type='pci' managed='yes'>
          <driver name='vfio'/>
          <source>
            <address domain='0x0000' bus='0x41' slot='0x00' function='0x0'/>
          </source>
          <address type='pci' domain='0x0000' bus='0x08' slot='0x00' function='0x0'/>
        </hostdev>

    Also is that change persistent in the xml ? or do i have to manually add it every time i restart the VM?

     

    Thank you so much for all your efforts

  2. On 10/7/2022 at 3:18 AM, ghost82 said:

    Add an hostdev block, set the source address of the device you want to passthrough, set the target address of the passed through device in the vm, attach the device in the target guest to bus 0 for machine type i440fx, attach the device to bus 0 --> x for machine type q35 (bus 0 in q35 is like a "built-in device", but this is not your case; for q35, if bus is different than 0 check that you have a pcie-root-port with the index number equal to that of the target bus).

     

    For a q35 machine the block will be something like this:
     

        <hostdev mode='subsystem' type='pci' managed='yes'>
          <driver name='vfio'/>
          <source>
            <address domain='0x0000' bus='0x41' slot='0x00' function='0x0'/>
          </source>
          <address type='pci' domain='0x0000' bus='0x08' slot='0x00' function='0x0'/>
        </hostdev>

     

    1. Your source address is 41:00.0 (bus=41, slot=0, function=0)

    2. Target address is 08:00.0 (bus=8, slot=0, function=0)

    3. Check that pcie-root-port with index=8 exists:

        <controller type='pci' index='8' model='pcie-root-port'>
          <model name='pcie-root-port'/>
          <target chassis='8' port='0xd' hotplug='off'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x7'/>
        </controller>

    -----

    For i440fx the block will be something like this:

     

        <hostdev mode='subsystem' type='pci' managed='yes'>
          <driver name='vfio'/>
          <source>
            <address domain='0x0000' bus='0x41' slot='0x00' function='0x0'/>
          </source>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x05' function='0x0'/>
        </hostdev>

     

    1. Your source address is 41:00.0 (bus=41, slot=0, function=0), the same obviously

    2. Target address is 00:05.0 (bus=0, slot=5, function=0): i440fx has only bus 0

     

    If you get errors like double address in use, or something similar, check that the target address is not already in use by something else, in this case change the bus number (for q35) or the slot number (for i440fx).

     

    The VM is a Q35 Machine but i am not sure on how to  3. Check that pcie-root-port with index=8 exists:

     

    Thank you

  3. 20 hours ago, ghost82 said:

    If you have issues just post the vm xml or better the full diagnostics.

    Here is the full VM xml

    Quote

    <?xml version='1.0' encoding='UTF-8'?>
    <domain type='kvm' id='2'>
      <name>Home Assistant</name>
      <uuid>5efce766-c82d-4abb-1d51-e82f1f7c5e4f</uuid>
      <description>Linux VM</description>
      <metadata>
        <vmtemplate xmlns="unraid" name="Linux" icon="default.png" os="linux"/>
      </metadata>
      <memory unit='KiB'>8388608</memory>
      <currentMemory unit='KiB'>8388608</currentMemory>
      <memoryBacking>
        <nosharepages/>
      </memoryBacking>
      <vcpu placement='static'>4</vcpu>
      <cputune>
        <vcpupin vcpu='0' cpuset='8'/>
        <vcpupin vcpu='1' cpuset='28'/>
        <vcpupin vcpu='2' cpuset='9'/>
        <vcpupin vcpu='3' cpuset='29'/>
      </cputune>
      <resource>
        <partition>/machine</partition>
      </resource>
      <os>
        <type arch='x86_64' machine='pc-q35-6.0'>hvm</type>
        <loader readonly='yes' type='pflash'>/usr/share/qemu/ovmf-x64/OVMF_CODE-pure-efi.fd</loader>
        <nvram>/etc/libvirt/qemu/nvram/5efce766-c82d-4abb-1d51-e82f1f7c5e4f_VARS-pure-efi.fd</nvram>
      </os>
      <features>
        <acpi/>
        <apic/>
      </features>
      <cpu mode='host-passthrough' check='none' migratable='on'>
        <topology sockets='1' dies='1' cores='2' threads='2'/>
        <cache mode='passthrough'/>
      </cpu>
      <clock offset='utc'>
        <timer name='rtc' tickpolicy='catchup'/>
        <timer name='pit' tickpolicy='delay'/>
        <timer name='hpet' present='no'/>
      </clock>
      <on_poweroff>destroy</on_poweroff>
      <on_reboot>restart</on_reboot>
      <on_crash>restart</on_crash>
      <devices>
        <emulator>/usr/local/sbin/qemu</emulator>
        <disk type='file' device='disk'>
          <driver name='qemu' type='qcow2' cache='writeback'/>
          <source file='/mnt/user/domains/HassOS/haos_ova-7.1.qcow2' index='1'/>
          <backingStore/>
          <target dev='hdc' bus='sata'/>
          <boot order='1'/>
          <alias name='sata0-0-2'/>
          <address type='drive' controller='0' bus='0' target='0' unit='2'/>
        </disk>
        <controller type='pci' index='0' model='pcie-root'>
          <alias name='pcie.0'/>
        </controller>
        <controller type='pci' index='1' model='pcie-root-port'>
          <model name='pcie-root-port'/>
          <target chassis='1' port='0x10'/>
          <alias name='pci.1'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0' multifunction='on'/>
        </controller>
        <controller type='pci' index='2' model='pcie-root-port'>
          <model name='pcie-root-port'/>
          <target chassis='2' port='0x11'/>
          <alias name='pci.2'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x1'/>
        </controller>
        <controller type='pci' index='3' model='pcie-root-port'>
          <model name='pcie-root-port'/>
          <target chassis='3' port='0x12'/>
          <alias name='pci.3'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x2'/>
        </controller>
        <controller type='pci' index='4' model='pcie-root-port'>
          <model name='pcie-root-port'/>
          <target chassis='4' port='0x13'/>
          <alias name='pci.4'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x3'/>
        </controller>
        <controller type='virtio-serial' index='0'>
          <alias name='virtio-serial0'/>
          <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/>
        </controller>
        <controller type='sata' index='0'>
          <alias name='ide'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/>
        </controller>
        <controller type='usb' index='0' model='qemu-xhci' ports='15'>
          <alias name='usb'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/>
        </controller>
        <interface type='bridge'>
          <mac address='52:54:00:a7:e7:9e'/>
          <source bridge='br0'/>
          <target dev='vnet1'/>
          <model type='virtio-net'/>
          <alias name='net0'/>
          <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/>
        </interface>
        <serial type='pty'>
          <source path='/dev/pts/1'/>
          <target type='isa-serial' port='0'>
            <model name='isa-serial'/>
          </target>
          <alias name='serial0'/>
        </serial>
        <console type='pty' tty='/dev/pts/1'>
          <source path='/dev/pts/1'/>
          <target type='serial' port='0'/>
          <alias name='serial0'/>
        </console>
        <channel type='unix'>
          <source mode='bind' path='/var/lib/libvirt/qemu/channel/target/domain-2-Home Assistant/org.qemu.guest_agent.0'/>
          <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/>
          <alias name='channel0'/>
          <address type='virtio-serial' controller='0' bus='0' port='1'/>
        </channel>
        <input type='tablet' bus='usb'>
          <alias name='input0'/>
          <address type='usb' bus='0' port='1'/>
        </input>
        <input type='mouse' bus='ps2'>
          <alias name='input1'/>
        </input>
        <input type='keyboard' bus='ps2'>
          <alias name='input2'/>
        </input>
        <graphics type='vnc' port='5900' autoport='yes' websocket='5700' listen='0.0.0.0' keymap='en-us'>
          <listen type='address' address='0.0.0.0'/>
        </graphics>
        <audio id='1' type='none'/>
        <video>
          <model type='qxl' ram='65536' vram='65536' vgamem='16384' heads='1' primary='yes'/>
          <alias name='video0'/>
          <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0'/>
        </video>
        <hostdev mode='subsystem' type='usb' managed='no'>
          <source>
            <vendor id='0x10d5'/>
            <product id='0x5000'/>
            <address bus='1' device='5'/>
          </source>
          <alias name='hostdev0'/>
          <address type='usb' bus='0' port='3'/>
        </hostdev>
        <hostdev mode='subsystem' type='pci' managed='yes'>
          <driver name='vfio'/>
          <source>
            <address domain='0x0000' bus='0x43' slot='0x00' function='0x0'/>
          </source>
          <alias name='hostdev1'/>
          <address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/>
        </hostdev>
        <memballoon model='none'/>
      </devices>
      <seclabel type='dynamic' model='dac' relabel='yes'>
        <label>+0:+100</label>
        <imagelabel>+0:+100</imagelabel>
      </seclabel>
    </domain>

     Here is also the full diagnostics zip

     

    Thank you thank you

    predator-diagnostics-20221008-1115.zip

  4. Hello everyone i am running 2 VMs on my unRAID server a Windows 10 VM and Home Assistant VM both running beautifully. I am also passing already 2 PICe devices (USB boards) to each of the VMs and an AMD GPU to the Windows VM.

    I bought a Coral Edge TPU to use with Frigate on Home Assistant which is a PCIe type card that i can see under Tools>System Devices on its own IOMMU group.

    I did bind the device to the VFIO and restarted the machine.

    In the VFIO-PCI Log there are no errors and it is successfully bound.

    However when I go to edit the VMs (any of them) there is no option listed under Other PCI devices for the Coral Edge TPU!

     

    What am I missing or doing wrong?? I have started the machine a couple of times to no avail

    Any help will be greatly appreciated

    Thank you

    Denis 

     

    vfio.jpg

    otherpcie.jpg

    logvfio.jpg

  5. Anyone has any experience with Nvidia GPU fan control on unRAID?? I have done some reasearch and my server is not headless i have a monitor connected to it and it is displaying unRAID GUI just fine also the GPU is not passed to any VM it is strictley for unRAID and docker use.

    The fan speed is running at 128% and the card is cold as ice. i have tried this commands:

     

    nvidia-xconfig --cool-bits=28 (or 4 or 31)

    nvidia-settings -V -c :0 -a [gpu:0]/GPUFanControlState=1 -a [fan:0]/GPUTargetFanSpeed=30

     

    that should do the trick but all i am getting is this:

     

    ERROR: The control display is undefined; please run `nvidia-settings --help` for usage information.

     

    which means that your computer does not have a screen attached to it and it needs a new X server to be written to run headless anything that i have tried to do that does not work and errors out.

    My card is a Nvidia Quadro P4000 and running the latest drivers and works fine other than the crazy fan speed

    any one else has any experience with this or is knowledgable enough to help please chime in and help me out

     

    thank you

  6. Anyone has any experience with Nvidia GPU fan control on unRAID?? I have done some reasearch and my server is not headless i have a monitor connected to it and it is displaying unRAID GUI just fine also the GPU is not passed to any VM it is strictley for unRAID and docker use.

    The fan speed is running at 128% and the card is cold as ice. i have tried this commands:

     

    nvidia-xconfig --cool-bits=28 (or 4 or 31)

    nvidia-settings -V -c :0 -a [gpu:0]/GPUFanControlState=1 -a [fan:0]/GPUTargetFanSpeed=30

     

    that should do the trick but all i am getting is this:

     

    ERROR: The control display is undefined; please run `nvidia-settings --help` for usage information.

     

    which means that your computer does not have a screen attached to it and it needs a new X server to be written to run headless anything that i have tried to do that does not work and errors out.

    My card is a Nvidia Quadro P4000 and running the latest drivers and works fine other than the crazy fan speed

    any one else has any experience with this or is knowledgable enough to help please chime in and help me out

     

    thank you

  7. I was trying to set up My Servers following SpaceInvader One video and everything was going fine untill it gave me a unraid.net subdomain https://ad4cce48c34e1d2d49e5fb8XXXXXXXXXXXXXXX.unraid.net/ 

     

    Now i can not reach my server even from inside my network using the internal ip address of 192.168.5.88, that has been static forever, because it tries the above https one 

    i was not able to change the 443 port forwarding rule and i do not have that port open on my router so the reaching form outside of the network is imposible also (althoughi tried opening that port and closing it back up) ...what do i do??

     

    Please help and thank you in advance

     

    By the way everything else works fine Plex, my VMs everything just can't reach the web interface

     

    Denis 

  8. EDIT: Never mind i got it ...to everyone else running into this; don't add the variable via SpaceInvaderOne style just read and carefully follow the instructions at the top 

     

    Hello all i am trying to install Unmanic and i am running into the following error when i click apply.

    "usr/bin/docker: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused "process_linux.go:449: container init caused "process_linux.go:432: running prestart hook 1 caused \"error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: device error: unknown device id: false\\n\""": unknown. "

     

    I am running Nvidia-unRAID version 6.9.0 beta_22 and i have an older Nvidia GeForce GTX980TI i am declaring the extra parameters and the NVIDIA_VISIBLE_DEVICES with the one Gpu UUID that i have but i get that error still. I use that gpu with handbrake and plex and even once with tdarr with no errors or problems everything works fine for those containers.

    what am i doing wrong ?? is my video card even capable of doing a decent job of converting or helping to h265 so i can save some space or not really, the cpu will be the one to do the work.

     

    thank you 

    nvidiaUnmanic.png

  9. Hey all as a result of a power outage and my UPS not configured right i had an ungraceful shut down and when i recovered both my docker engine and my vms would not start.

    With the help of a couple of gentleman from the forums I did get my Docker service up and running but i can't get my VM ...i do not have a copy of the xml file but i do have a copy of my libvirt image and my vdisk 

    How do i get my old vm back?? I tried to mess around with it looking at blog posts but now, although i have turned off and back on VM settings, i keep getting the Libvirt Service failed to start.

    I am attaching a copy of my diagnostic file 

    Any help is greatly appreciated

    Thank you in advance

    Denis

    tower-diagnostics-20200905-2011.zip

  10. Hey thank you for the quick response...

    1 hour ago, trurl said:

    That seems to indicate that the version on cache is the currently used one, as it should be. You can delete the system folder on disk1

    That seems to be the copy of my old VM that i lost how do i use that to get my VM back .....i have been browsing all over but i keep reading about the old xml file that i dont have

  11. Hello Gentleman,

    Well i was up and running for a couple of days and than back to square one. I tried to restart the PiHole docker container and  got error 403 and it says can not write to docker image either full or corrupted. And the damn log is still spammed by the ACPI error although i have the  rmmod acpi_power_meter added to the go file. So i restarted the tower and the docker came back up so i removed the pi hole (i wanted to anyhow) than i executed in the terminal rmmod acpi_power_meter again and the ACPI errors stopped. My array does not automatically start up after a reboot and as a result the docker service is not up and running so the command rmmod acpi_power_meter has to be executed after Docker service has started for it to work  ...so how do we do that??

     

    Cache gets full pretty quickly when downloading loads at the same time and/or converting with handbreak/tdarr. What do i have to change to always have room in the cache ( i have the Mover scheduled to run hourly) ... I blamed PiHole for the docker hiccup this time but i think previously it was either handbreak or tdarr that caused problems with the cache and docker .... also Sonarr/Radarr move about 90% of the Completed downloads from the /cache/downloads/completed folder i have no clue why they wont move 100% of them to Media/movies share

     

    On 8/26/2020 at 9:49 AM, trurl said:

    From the command line, what do you get with these?

    
    ls -lah /mnt/cache/system/libvirt
    
    ls -lah /mnt/disk1/system/libvirt

     

    Here is what i am getting for the above commands  .... looks like the same file but different sizes...by the way i used to have a VM that disappeared at the same time with my dockers the first time and i tried to recreate it again.

     

    Thank you

    Denis

     

     

    disk1libvirt.png

    cachelibvirt.png

  12. Good evening gentlemen 

    On 8/24/2020 at 3:49 AM, johnnie.black said:

    Cache is balanced now but the syslog is being spammed with ACPI errors, see here for how to fix then reboot and post new diags.

    I did the modification of /boot/config/go to include the rmmod acpi_power_meter and rebooted when i rebooted the Docker service was set to start and before i  started the array i checked there were no ACPI errors when i started the array and the Docker service started automatically i saw the ACPI errors  started surfacing up again. How is that correlated somehow cause i thought we are dealing with hardware os level stuff not Docker....... and i made sure that it had saved the change on the /boot/config/go file to include the rmmod.

    So in other words the spamming error still remains even though the config is modified to kill the acpi_power_metering...how??? it is beyond my understating ability LOL

     

    On 8/23/2020 at 9:24 PM, trurl said:

    Do you know how to examine the disks?

     

    Do you have any VMs?

    i did examine the disk and yes there is a vm i am trying to run on disk1 system folder...is it supposed to be there?? how do we move that an others it in order for the spin downs to occur .

     

    thank you again waiting for the next troubleshooting steps 

    Denis

    ps diagnostics zip included 

    disk1.png

    disk1a.png

    bootconfig.png

    tower-diagnostics-20200825-2113.zip

  13. Ok gentleman i removed some files from the Cache and run the balance and got a successful message although the empty space seems to have decreased

    (I am having a hard time comprehending this) i have turned off everything VMs and Dockers.

    I run fix common problems and this is what i got .....that the docker image might be full or corrupted. I did try to turn the Docker service on briefly and i got the Docker Service failed to start so it turned it back off.

    What do we do next?? Should i delete the Docker image again?? What did this fix because i have a feeling that if i delete the image and start installing the apps again the same thing will happen all over. 

    I am attaching the new Diagnostic zip file.

     

    Thank you as always 

    Denis

    btrfsCache.png

    FixCommonProb.png

    tower-diagnostics-20200823-1144.zip

  14. @johnnie.black @trurl I greatly appreciate your feedback as i am trying to better understand unRAID

     

    I followed the guide given on the above link for balancing the cache drive with the btrfs balance start -dusage=75 /mnt/cache command and it would not let me i tried with -dusage=1 and it worked all the way up till the whopping 4 :) but when i did 5 it says no space available.

     

    So i am trying to remove some files from the cache disk from the /downloads/completed folder that Sonar has not moved (i have no idea why it moves the majority but not all) and see if that will solve some of the problem however as stated on my original post the system reports a 107Gb free space on Cache.

    Truth be told i am a bit unsure on what the step by step path forward for a solution to this is??

    What would one do Steps 1,2,3 ...x

     

    On 8/21/2020 at 7:53 AM, trurl said:

    Why do you have 50G allocated to docker.img? Have you had problems filling it? 20G should be more than enough, but I see you are already using 18G. I suspect you have some application misconfigured and it is writing to a path that isn't mapped.

    I suspect that the instance of Tdarr is not configured right but i never used it however the first time i had to increase from 20 to 50g was right after i installed that app when i stopped it the issue did not replicate.  

     

    On 8/21/2020 at 7:53 AM, trurl said:

    You shouldn't run for long on batteries. The point of having batteries is so you can cleanly shutdown, not so you can keep running. Is your UPS compatible with the APCUPSd builtin to Unraid?

    The ups i had was not APCUPSd compatible (it would recognize the UPS but only 2 parameters would report and incorrectly) and i am changing that now with a genuine APC smart and on top of all that i was not home when it happened and it lasted for 4+ hrs.

    Question: Will the built in UPS feature take care of the shutdown process automatically if connected to a genuine APC ups??

    On 8/21/2020 at 7:53 AM, trurl said:

    Also, your system share has files on disk1 instead of all on cache. Possibly this happened when you recreated docker.img while still having cache problems, or maybe you enabled dockers / VMs before you installed cache. Your dockers/VMs will keep those files open, so array disks can't spin down, and they will have their performance impacted by slower parity.

    I have had a cache drive from the beginning when i set up the system however i had the System share to use Cache: Preferred instead of Cache: Yes...

    Will a simple change of that setting fix that issue or do i have to do something special now??

     

    Looking forward to your help and guidance on these issues and quick guide on how to get this up and running the soon

     

    Thanks again

    Denis