Execut1ve

Members
  • Posts

    29
  • Joined

  • Last visited

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

Execut1ve's Achievements

Newbie

Newbie (1/14)

2

Reputation

  1. It seems it was too good to be true... pausing behavior has returned. Based on the syslog, it does seem to be related to an issue with one of the graphics cards: Oct 15 10:12:11 RemoteUnraid kernel: pcieport 0000:00:07.0: AER: Multiple Uncorrected (Fatal) error received: 0000:00:00.0 Oct 15 10:12:11 RemoteUnraid kernel: vfio-pci 0000:06:00.0: AER: PCIe Bus Error: severity=Uncorrected (Fatal), type=Inaccessible, (Unregistered Agent ID) Oct 15 10:12:12 RemoteUnraid kernel: pcieport 0000:00:07.0: AER: Root Port link has been reset Oct 15 10:12:12 RemoteUnraid kernel: pcieport 0000:00:07.0: AER: device recovery successful I'm going to try reseating all the cards and risers and see if that helps any
  2. I actually got the VM to work without pausing by adding a USB keyboard and mouse and passing them through to the VM. I have no idea why that would work, but the VM did originally have those items. I had removed them when I physically relocated my server but now the VM is working after re-adding them. I typically access the VM via VNC. I usually don't have any issues with it, except occasionally the disappearing mouse cursor.
  3. I just tried extending the vdisk... I was able to add space but the VM still pauses. If I try to force it to resume, I get Execution error internal error: unable to execute QEMU command 'cont': Resetting the Virtual Machine is required
  4. Hibernation is turned off, and I believe sleep is as well - however I can't get the VM to run long enough to check before it pauses.
  5. My Windows 10 VM has lately started pausing itself for no apparent reason. Diagnostics attached. From some cursory research, it looks like it may be related to drive space, but I have only one drive in my system (aside from the parity drive) and it shows plenty of free space so I'm not sure how to resolve. Any help appreciated! remoteunraid-diagnostics-20211013-1051.zip
  6. Is anyone having issues with the latest update of the MobiusNine Folding@Home docker container? After updating, Folding@Home is suddenly not detecting any CUDA or OpenCL and all my slots are showing as disabled.
  7. Not sure where the best place would be to ask this, so apologies if this doesn't belong here - Can anyone point me to a way to automate logging of CPU temp into a .csv file or something for later analysis? Thanks!
  8. After some informal experimentation, I'm not seeing much difference (if any) in my total PPD between allocating the container 4 cores (with nothing on the HTs) VS allocating 2 cores with their 2 HTs. For reference I'm folding on 4 GPUs: 3 of the Zotac 1060 mining variants and 1 GTX960. They are connected to the mainboard via powered PCIE riser cables. Two of them are in x8 slots and two are in x4 slots. All the PCIE slots are Gen 2. The computer is a PowerEdge R710 server with dual Xeon X5690 processors. I'm averaging 800k-1M total PPD, with each card sitting in the 200k-250k range. I don't notice any substantial difference between the cards on the x4 slots vs the x8 slots. Can anyone else with a hyperthreaded CPU offer any observations?
  9. Hm, I wonder if I'd notice a hit to folding performance if I assigned the container 2 cores and 2 hyperthreads instead of 4 cores? Time for some experimentation!
  10. I've been using the docker container to fold with 4 GPUs / no CPU. Everything seem to be working well but I've noticed that the container seems to use a CPU core for each GPU slot, and each CPU core it uses is pinned at 100% utilization. Is anyone else getting similar behavior? I realize the GPUs have to be fed data to fold, but it seems like that shouldn't take up 100% of a core. The CPUs are Xeon x5690s, so not exactly new, but not slouches either. Can anyone offer any thoughts? Am I misunderstanding something in how all this works?
  11. In case anyone is curious how this turned out, I did some research into the Intel 5520 chipset which is on the R710. The architecture seems to use 3 of what Intel calls Quick Path Interconnects to connect the 2 processors together and each processor to the I/O hub. So unlike some other architectures (such as the one in Gridrunner's tutorial) the PCIe slots seem to connect directly to both processors rather than just one or the other.
  12. Hello all, I recently followed Gridrunner's excellent tutorial here on how to use the lstopo command to optimize PCI slot assignment, CPU core assignment, etc for best performance of VMs in a multi CPU setup. I am using a PowerEdge R710, which is a dual CPU board, with 2x Xeon X5690 installed. I was expecting to see output similar to the example in the video, though with 2 CPUs instead of 4, and I'd have to match up which PCI slot is connected to which CPU. Instead this is what I got: There are a couple things about this that don't make sense to me: -why are the numbers of the CPU cores all wonky? They are each 6 cores (12 threads), I'd expect to see them numbered 0-5 and 6-11 or something like that -why do all the devices seem to be connected to the same CPU? My understanding was each CPU has its own memory and some (but not all) of the PCI lanes, is this not correct? -what do the numbers (eg 1.0, 2.0, 0.5) on some of the devices mean?
  13. I just use Windows 10. Performance is more than adequate for games at 1080p with high quality settings. I don't do any FPS games though, so I can't really comment on whether the latency would be tolerable for that
  14. I like a lot of cache because I have a couple daily driver type VMs that live entirely on cache, including a hundred G or so apiece for games, frequently used programs etc. Each VM also has storage space on the array for bulk items that don't need to be fast. The general thinking is your cache should be large enough to comfortably hold an entire day's worth of new writes to the array. Then, when the Mover is run, it all gets written to the array and you start over fresh the next day. My understanding is if the cache fills up and you need to do still more writes, you'll write directly to the array and lose any of the speed gains you have from the faster cache. So you should determine your cache size based on how much data you expect to write in a day to your array.
  15. 3, actually. 2 250G and 1 500G SSD for a total of 500G usable cache space. My next upgrade will most like be swapping in some larger SSDs and probably going down to 2 drives in the pool, that way I have room for another big spinning array drive if needed