Jump to content

ogi

Members
  • Posts

    288
  • Joined

  • Last visited

Posts posted by ogi

  1. 5 hours ago, Falcosc said:

    Check first what happens with the CPU state after enabling L0s before investigating this. And then compare the power draw without this card to see if it is worth to hunt for.

     

    With Ubunutu I was able to reach C8 state in powertop which resulted in a major power draw change of 4W.

     

    And don't forget to execute autotune with powertop

     

    I don't think my E5-2680v2 processors can reach C8 state, would love to go into C7 tho!

     

    image.thumb.png.59cc985d6dc021408a0f4940603adff1.png

  2. 12 minutes ago, Falcosc said:

    @ogiif Ubuntu works, I encourage you to create a new bug for this issue. You can have a look at

    to see which kind of information are needed for evidence.

     

    But if Ubuntu doesn't work for your devices, then you have a real setup issue. There could a specific hardware combination which does prevent it or a configuration issue.

    For example, your LSI SAS2308 PCI-Express Fusion-MPT SAS-2 doesn't have L1 support at all (LnkCap). The 4 Port SATA on the other hand have it. Have an eye on that, this one device could prevent your CPU from sleep. And even if 9 of 10 links are sleeping, you won't see much power savings as long as a single device keeps your CPU busy on the bus.

     

    For that reason, I recommend checking power consumption on Ubuntu. You don't need to spend time on ASPM for the supported devices, as long as you have at least one thing which will not support it. I don't know if L0s is enough to see significant changes in the power consumption, I only know about the huge effect of a proper sleeping CPU after having everything in the deeper L1 sleep.

     

    I'll give the LiveUSB a try in a bit and report back (might be a day or so, the garage where my server resides is absolutely cooking right now, and I don't really want to spend much time there).

     

    Regarding the HBA, is there a known HBA that can be flashed to IT mode that supports L1 support?

  3. 2 minutes ago, Falcosc said:

    Have you already done the usual solutions like bios setup and kernel boot options? If your system is setup incorrectly, it will not work.

     

    If you have everything, then you should check Ubuntu first, because Ubuntu does not have the unknown ASPM bug which Unraid has for my hardware.

    If you can get ASPM enabled on Ubuntu without hacks, then we will use your result to push forward. It will provide more evidence to help Unraid team to identify the cause of the issue.

     

    And I would only recommend the ASPM hack (forced activation, which does bypass any compatibility checks) if you did confirm that other operating systems can enable it, then bypassing logic is a bit less dangerous.

     

    I set my BIOS ASPM option to "auto", here is the blurb in the manual about the options.  I didn't actually see the Force L0 option.

     

    image.thumb.png.41073835155644aa44b331f238f5f6a3.png

     

    I also added pcie_aspm=force to the list of options in my /boot/syslinux/syslinux.cfg file

     

    root@Tower:~# cat /boot/syslinux/syslinux.cfg
    default menu.c32
    menu title Lime Technology, Inc.
    prompt 0
    timeout 50
    label Unraid OS
      menu default
      kernel /bzimage
      append intel_iommu=on rd.driver.pre=vfio-pci video=vesafb:off,efifb:off isolcpus=4-9,24-29 vfio_iommu_type1.allow_unsafe_interrupts=1 initrd=/bzroot pcie_aspm=force
    label Unraid OS GUI Mode
      kernel /bzimage
      append isolcpus=4-9,24-29 vfio_iommu_type1.allow_unsafe_interrupts=1 initrd=/bzroot,/bzroot-gui
    label Unraid OS Safe Mode (no plugins, no GUI)
      kernel /bzimage
      append initrd=/bzroot unraidsafemode
    label Unraid OS GUI Safe Mode (no plugins)
      kernel /bzimage
      append initrd=/bzroot,/bzroot-gui unraidsafemode
    label Memtest86+
      kernel /memtest

     

     

    Before the above two options, all entries when I did lspci showed ASPM disabled, with the above changes, only a handful of entries (primarily PCI bridges and my HBA) show ASPM as being disabled.

     

    I have no problem trying ubuntu; server is a bit tough to access but I'll make a go of it.  Would the idea be to boot off of a liveUSB and then check ASPM status?

  4. Sorry to bring a thread back from the dead; I decided that I wanted to spend time trying to minimize idle power-usage on my Supermicro Server (X9DRi-LN4F+ motherboard).

     

    I followed most of the advice here; have a handful of devices that ASPM is still disabled for here; hoping I could get some suggestions on how to address them:

     

    root@Tower:~# lspci -vv | awk '/ASPM/{print $0}' RS= | grep --color -P '(^[a-z0-9:.]+|ASPM )'
    00:00.0 Host bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 DMI2 (rev 04)
                    LnkCap: Port #0, Speed 2.5GT/s, Width x4, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk-
    00:01.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 1a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 2.5GT/s, Width x4, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk-
    00:01.1 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 1b (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x4, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk+
    00:02.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 2a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    00:03.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 3a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    00:11.0 PCI bridge: Intel Corporation C600/X79 series chipset PCI Express Virtual Root Port (rev 06) (prog-if 00 [Normal decode])
                    LnkCap: Port #17, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us
                    LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    00:1c.0 PCI bridge: Intel Corporation C600/X79 series chipset PCI Express Root Port 1 (rev b6) (prog-if 00 [Normal decode])
                    LnkCap: Port #1, Speed 5GT/s, Width x4, ASPM L1, Exit Latency L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    pcilib: sysfs_read_vpd: read failed: No such device
    02:00.0 Serial Attached SCSI controller: Broadcom / LSI SAS2308 PCI-Express Fusion-MPT SAS-2 (rev 05)
                    LnkCap: Port #0, Speed 8GT/s, Width x8, ASPM L0s, Exit Latency L0s <64ns
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk+
    03:00.0 VGA compatible controller: NVIDIA Corporation TU104GL [Quadro RTX 4000] (rev a1) (prog-if 00 [VGA controller])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <512ns, L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    03:00.1 Audio device: NVIDIA Corporation TU104 HD Audio Controller (rev a1)
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <512ns, L1 <4us
                    LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    03:00.2 USB controller: NVIDIA Corporation TU104 USB 3.1 Host Controller (rev a1) (prog-if 30 [XHCI])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <512ns, L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    03:00.3 Serial bus controller: NVIDIA Corporation TU104 USB Type-C UCSI Controller (rev a1)
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <512ns, L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    04:00.0 VGA compatible controller: NVIDIA Corporation GP106GL [Quadro P2000] (rev a1) (prog-if 00 [VGA controller])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    04:00.1 Audio device: NVIDIA Corporation GP106 High Definition Audio Controller (rev a1)
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    05:00.0 Serial Attached SCSI controller: Intel Corporation C602 chipset 4-Port SATA Storage Control Unit (rev 06)
                    LnkCap: Port #0, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us
                    LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    06:00.0 Ethernet controller: Intel Corporation I350 Gigabit Network Connection (rev 01)
                    LnkCap: Port #0, Speed 5GT/s, Width x4, ASPM L0s L1, Exit Latency L0s <4us, L1 <32us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    06:00.1 Ethernet controller: Intel Corporation I350 Gigabit Network Connection (rev 01)
                    LnkCap: Port #0, Speed 5GT/s, Width x4, ASPM L0s L1, Exit Latency L0s <4us, L1 <32us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    06:00.2 Ethernet controller: Intel Corporation I350 Gigabit Network Connection (rev 01)
                    LnkCap: Port #0, Speed 5GT/s, Width x4, ASPM L0s L1, Exit Latency L0s <4us, L1 <32us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    06:00.3 Ethernet controller: Intel Corporation I350 Gigabit Network Connection (rev 01)
                    LnkCap: Port #0, Speed 5GT/s, Width x4, ASPM L0s L1, Exit Latency L0s <4us, L1 <32us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    80:00.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port in DMI2 Mode (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #8, Speed 2.5GT/s, Width x48, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk-
    80:01.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 1a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x8, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    80:02.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 2a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk-
    80:03.0 PCI bridge: Intel Corporation Xeon E7 v2/Xeon E5 v2/Core i7 PCI Express Root Port 3a (rev 04) (prog-if 00 [Normal decode])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L1, Exit Latency L1 <16us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    82:00.0 Non-Volatile memory controller: Intel Corporation SSD 660P Series (rev 03) (prog-if 02 [NVM Express])
                    LnkCap: Port #0, Speed 8GT/s, Width x4, ASPM L1, Exit Latency L1 <8us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    84:00.0 VGA compatible controller: NVIDIA Corporation TU104GL [Quadro RTX 4000] (rev a1) (prog-if 00 [VGA controller])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    84:00.1 Audio device: NVIDIA Corporation TU104 HD Audio Controller (rev a1)
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    84:00.2 USB controller: NVIDIA Corporation TU104 USB 3.1 Host Controller (rev a1) (prog-if 30 [XHCI])
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    84:00.3 Serial bus controller: NVIDIA Corporation TU104 USB Type-C UCSI Controller (rev a1)
                    LnkCap: Port #0, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <1us, L1 <4us
                    LnkCtl: ASPM L1 Enabled; RCB 64 bytes, Disabled- CommClk+
    

     

    Is this realistically the best that can be done with respect to ASPM enabling or is there further room for improvement?

  5. 1 hour ago, neural said:

    Hi,

     

    May I ask what GPU or cards you have installed? 

     

    We are trying to troubleshoot this same error with a K80, Quadro 4000 and one LSI HBA Card.

     

    The system fails when we start to add any drives to the system or any real use of Unraid. The system will just die. 

    I only had this issue with my GTX 670.  I have had GPU pass through work with the Quadro RTX 4000 and a RTX 3080 FE.

  6. 3 minutes ago, mgutt said:

    Only to be sure: The random.file was written to the disk where the movie of the kids was located, too?

     

     

    Sorry for the confusion; random.file was written to the same disk that a very high bitrate movie was streaming from (disk 18).  I'm not sure which disk random.file was written to when I aimed it at /mnt/user/movies/random.file (haven't looked for it yet).  

     

    The kids movie, I'm fairly certain is on another disk in the array.  All operations went smoothly so now I am having trouble even replicating the issue I began with (where the movie would freeze for 5-10 seconds, and there would be no data coming in via the network during that time).  

  7. Thanks for chiming in @mgutt  It's been running for ~10 minutes w/o issue; have to go back to monitoring the kids.... will edit this post later after the dd command returns.

     

    EDIT: kids now want to watch a movie too; so this will be a real stress test haha.

     

    EDIT2:  dd command finished

     

    root@Tower:~# dd if=/dev/random of=/mnt/user/movies/random.file bs=128k iflag=count_bytes count=128GB
    976562+1 records in
    976562+1 records out
    128000000000 bytes (128 GB, 119 GiB) copied, 1954.55 s, 65.5 MB/s

     

    Everything seems to be working remarkably well ...two movies playing (one on Plex via transcoding, BTW thanks for your in-memory decoding guide)... I guess whatever I have had going on before, it's no longer an issue? ...maybe I should test some CPU intensive tasks too...but I can do that later.

     

    EDIT 3: Just wanted to say thanks again for your debugging steps.

  8. It's running now; I'm getting occasional hangups in the stream; here is a screenshot when during one of those freezes... for the most part its playing well.

    duringFreeze.thumb.png.de38696cda527c172a194a44aacf3e70.png

     

    The `dd` command finished

     

    root@Tower:~# dd if=/dev/random of=/mnt/disk18/movies/random.file bs=128k iflag=count_bytes count=128GB
    976562+1 records in
    976562+1 records out
    128000000000 bytes (128 GB, 119 GiB) copied, 2241.48 s, 57.1 MB/s
    root@Tower:~#

     

    Given the movie (mostly) played w/o interruption, not sure what to make of it.

  9. I have a 2x  E5-2680 v2 CPUs; I use dual parity drives inside a Supermicro CSE-846 chassis.  My situation I don't think is Plex specific; and unfortunately I couldn't replicate last night while watching a high bitrate video (video is 82615 kb/s), but I did just upgrade from unraid 6.8.3 to 6.9.1.

     

    When streaming very high bitrate video (on VLC for examples) periodically the video stream will outright stop for ~10 seconds... and then resume.  I have assumed this was due to some other write operation to the array; but I cannot be certain.  I should point out things did get better over NFS vs. SMB; but I suspect that's just because the faster file transfer allowed for faster buffering (getting NFS mounts on macOS took me a while to sort out, I should probably post a guide about that!)

     

    I'll try and monitor things in iotop and the "System Stats" page on the webui and see if I can get a better idea of what else is happening when this occurs.

     

    As your issue involved Plex transcoding, and mine seems to have to do with just general file read operations, I'll create a new post if I'm able to reproduce more consistently.

     

    Thanks for chiming in!

  10. 1 hour ago, Kvo1087 said:

    Hi Guys! i have been fighting with my Supermicro X9DRi-LN4F+ for a couple days now and finally got IPMITool to work after a bios and firmware update. i tried to follow the instructions on the first page here setting my lower thresholds down around 100-200 for my noctua fans and turning the fan mode to full speed on the motherboard to run this pluggin. unfortunately right now the fans still run full speed all the time and if i change the fan mode it goes back to cycling from low rpm to max speed every few seconds. any help would be greatly appreciated. :)

     

    i think my main issue might be that i cant get the virtual machine overide to work. it defaults to no and if i change it to yes everything is greyed out and i cant fill it in. then i hit apply and it goes back to "no" on the setting. I'll attach a picture for reference

     

    If i could get the fan speed minimum and maximum settings to work in the fan control section that would be my greatest desire. i currently have it set based on cpu temp with max fan speed 80% and min 35%. they are definitely going full speed...

    virtual machine overide.JPG

    I have this same motherboard; ...that's curious why everything is grayed out...I had to specify a network connection and have it register "connection successful" before even attempting to do anything else... here's a screenshot of my settings.

     

    image.thumb.png.aa9e2afec9d694083082d6a94821c43f.png

  11. On 5/17/2020 at 4:07 PM, phrozen087 said:

    Were you able to get this to work? I have 2 Samsung 970 NVME drives that consistently run in the 40-45C range as Unassigned Devices to hold my VM images, so I manually went in and set all of my spinning disks to have have 40C warning/ 45C critical and then changed my global disk setting (Settings->Disk Settings-> Default warning disk temp and Default critical disk temp) to 65C warning and 70C critical.

     

    I still see warnings whenever I log into my server that one of the NVMEs was running in the 45-50C range, which is pretty common during VM disk images backups.  I've tried changing this up and down thinking maybe its not sticking, but it seems like the temps for my unassigned devices are stuck at the default temperatures from when unraid installed and aren't observing the values in this setting. Wasn't sure if I was doing something wrong here, or if there was an easy fix I was just missing?

    I just implemented the suggested work around, I set the temperature limits for the hottest running drive, and set temperature limits for my array disks individually. 

  12. you can definitely modify the thresholds from within the app; you need to go to config settings and select sensors from the drop down.  Careful removing a fan, you cannot get it back (unless you factory reset the BMC controller). On my x9 board, I can only set the thresholds on increments of 75.  For example I can set a threshold at 750, or lower it to 675, but I cannot set it to 700 even.

  13. 10 minutes ago, johnnie.black said:

    It's been requested before, for now the usual workaround is to set the higher temps in the general settings, then set lower temps for each array device.

    Figured this had been requested before, with 167 pages in the thread, and the search results for "temperature" was still at least 3 pages.... was hard to determine what the current status is.

     

    Thanks for the suggested workaround!

     

  14. hi dlandon,

     

    Thanks for maintaining this plugin, it's amazing the functionality it provides!

     

    I didn't see if there was a feature request for this already, but I was hoping you could offer functionality for different temperature thresholds?  I have a SSD as an unassigned device, which runs hotter than my spinning disks, and often get numerous warnings about excessive temperature and such.

     

    Thanks again for your work on this plugin, it's been an absolute life saver!

    • Like 1
  15. 2 minutes ago, acbaldwi said:

    thanks I appreciate it. Did you adjust the fan percentages at all on the gui?

    I did, those settings should be captured in fan.cfg, but aren't easy to read, here a screenshot from my GUI

     

    image.png.ad69dbbed3247323c52bb5279b31cb99.png

     

    These settings have kept things pretty quiet for me hard drives never get more than 45C during a party check.  Next time I power off the machine, I'm going to install the Noctua low-noise adapters for my CPU fans, and the 2 fans connected to FAN5/6 (but I haven't done that yet).

  16. 1 hour ago, acbaldwi said:

    Ogi i'm running into similar type usses can you please upload your cfg files and what your percentages are i'd like to see how mine differ maybe we can find a correlation...

     

    Thanks,
    Aaron

    Attached are my fan.cfg and ipmi-sensors.config files which reside in /boot/config/plugins/ipmi

     

    Keep in mind the settings there are for the fans I'm using, in the configuration I have.  I have Noctua iPPC 3000 fans connected to FAN1/2/3, and I have some Noctua 80mm fans connected to FAN5/6.  The CPU fans (92mm, can't remember the model) I have connected in a y-splitter connected to FANA.

     

    One of the things I noticed is that once you remove a fan setting from the config, there is no way to add it back, until you restore the IPMI controller to factory settings, so careful removing fan settings from the sensors config!

    ipmi-sensors.config fan.cfg

  17. Thanks for chiming in @jonpI didn't for you to answer support tickets on a weekend!  I bought this GPU ages ago, to say I got my money's worth out of it would be an understatement, I'm okay with retiring it.

     

    I suppose before I buy a new GPU, I should re-purpose the P2000 GPU into one of those slots to make sure passthrough works as intended and that it is in-fact this GPU that is causing the issue, I'll update this thread if I turn up anything else of interest.

    • Like 1
  18. Hi,

     

    I do indeed have 2 GPUs, the monitor output I described above was through the onboard VGA connector.  Before I describe the PCIe layout, probably best you look at the photo of the motherboard here: 

     

    https://www.supermicro.com/products/motherboard/Xeon/C600/X9DRi-LN4F_.cfm

     

    The first slot, closest to the CPUs is a 4x slot in a 8x connector (occupied with a NVMe adapter for my cache drive).  The slot furthest from the CPUs holds a Quadro P2000 GPU, which I use primarily for plex transcoding (running in Docker).  That slot furthest from the CPU is up against the chassis, so there is no way a 2x width card can fit.  Adjacent to that slot, is an actual 8X connector which I have my HBA attached to.  That leaves the 3 PCIe 16x slots, 2nd from the CPU to the 4th from the CPU.  As the GTX 670 is a dual slot width card, I can only use two of those slots, and I've tried both at this point.

     

    Thanks for chiming in, I really do appreciate a second set of eyes on this issue!

  19. Well, when the GPU goes invisible the way I described earlier, all I need to do is plug it into my desktop, power it up, then put it back into the server, and wha-la it's visible again.

     

    Anyway, I trie adding the following options when booting, but still with the same result.

    append isolcpus=16-19,36-39 pcie_acs_override=downstream,multifunction intel_iommu=on rd.driver.pre=vfio-pci video=vesafb:off,efifb:off initrd=/bzroot

    This time I had dmesg -wH running, I don't think it gave me any more meaningful information, but I'll post the screenshot regardless

    598519868_ScreenShot2020-03-14at8_56_37PM.thumb.png.67b307010a31cfce3eb33da5d09792e9.png

    At this point, I'm just starting to suspect that this GPU just plain won't work with passthrough.  I'd certainly welcome other things to try if anyone has other suggestions.

     

  20. Another oddity I'm discovering is that sometimes the GTX 670 is not even listed in the Tools -> System Devices, or visible in `lspci`

    root@Tower:/boot/config# lspci | grep 670
    root@Tower:/boot/config#

    On reboot the GPU is usually shown, but I've had to reboot on a few occasions now to ensure the device is visible in system devices..... i suppose i should try starting the VM again like this without the device being visible in unraid....server is being utilized somewhat heavily right now so this will have to wait :(

  21. From my googling, it sounds like 600 series cards, could be compatible, but generally weren't initially.  Manufacturers were distributing UEFI capable vBIOSs on request through the forums from the looks of things.  I confirmed in my desktop via GPU-Z, that my Gigabyte GTX 670 did not have UEFI capability.

     

    I should note that before I flashed the BIOS, no matter what configuration I had in the VM setup, i would always get Error Code 43; it wasn't until i read the bit on the wiki that OVMF devices must support UEFI booting.  I tried a SeaBios config, but could never get the VM to start.

  22. 4 minutes ago, bastl said:

    Where did you get the information to flash the BIOS directly to your card? NO ONE in the forums will ever give you that advice!!! Selecting the downloaded BIOS in the VM settings template is all you have to do. I never heared someone flashed the BIOS directly to the card for a GPU passthrough to work. Let's hope you not bricked your card. 

    Card is not bricked, it works fine on the desktop (that's where I did the flashing), and where I verified the card had UEFI capability.  The vbios I flashed on the card is meant for my model card, the difference it had vs. the vbios on my card is the UEFI capability, which is needed.  Here is the post with another user that discovered the same thing:

     

     

×
×
  • Create New...