WashingtonMatt

Members
  • Posts

    104
  • Joined

  • Last visited

Everything posted by WashingtonMatt

  1. Just to follow-up, I have recreated Libvirt.img and all errors have cleared up. Lost one VM to corruption, but was able to recover my important ones. I'm still feeling uneasy about the original issue. At this point I'm assuming it was due to not having Scrub scheduled, and if that is the case, shouldn't it be enabled by default? I could have sworn I had a dynamix plugin for that at some point. I did upgrade cache drives from mirrored 500GB to mirrored 2TB a few months back, had plenty of free space. And more recently upgraded from 6.11 to 6.12. If anyone can offer any insight as to what might have happened, I'd appreciate it. Just want to make sure I'm covering the bases now.
  2. So when the original event occurred the cache pool had gone read only. After some reading here, even though the UI showed I had tons of free space, I deleted a large test VM I didn't care about and then ran a scrub. The scrub reported it made a bunch of corrections(posted above). The errors seemed to stop after that. Now I'm getting the loop3 errors when I try to start some of the VM's. GUI appears to be complaining about TPM. Anyway, I disabled Docker and VM and ran a scrub again which came back clean. Re-enabled Docker and VM. Tried to start a problem VM and still get an error. unraid-diagnostics-20231127-0932.zip
  3. I deleted a large vm disk image I didn't care about, then ran a scrub which reported a ton of corrections. My VM disk images all seem to be corrupted. I was able to repair my critical VM enough that I can focus on it later. Are there BTRFS options that may help save my VMs? edit: and errors seemed to have stopped, except for the loop3 error, which I think occurred when I attempted to start a VM.
  4. Does this mean I've got a bad drive, or is it some other BTRFS unhappiness?
  5. Woke up to non-responsive dockers and VM's. I'm seeing a ton of cache pool disk errors. I have two mirrored nvme drives. I tried to copy a VM disk image to the array and the copy failed, so now I'm in a frustrated panic and not thinking clearly. I think I might have rebooted before I captured the first diagnostics. After rebooting I did recreate the docker.img since it was clearly corrupted, then rebuilt some dockers. Now I thinking even that was a bad idea. Really hoping we can save my cache drive data. Please assist. unraid-diagnostics-20231124-0533.zip unraid-diagnostics-20231124-0736.zip
  6. Thanks, with those changes I am able to play Fortnite again if EAC loads. Did not work for the new Roblox Hyperion anti-cheat. Finally got a great deal on some NVMe M.2 drives and upgraded my cache drives. Now I've got some space to build some testing VMs.
  7. Well, looks like Fortnite on VM is completely done now. Easy Anti-Cheat is now blocking Virtual Machines as well. My kids can no longer play Roblox either. This worked so well for so long, so frustrating. No way in hell I'm building gaming rigs for the whole family. The future is in VM's anyway, these companies are going to have to face that fact eventually. I've looked into virtual machine obfuscation a bit, and it's definitely not easy. If anyone comes across a good guide, please share.
  8. No solutions yet that I'm aware of. I've given up trying to workaround it. It's clearly above my pay grade. I'm still having luck playing when Easy Anti-Cheat loads. When when Battleye loads, it seems to be a minimum of 24 hours before EAC has a chance at loading again. If they keep it up, I might have to mow the lawn or something...
  9. No solution yet. I did find out that Fortnite uses two anti-cheat applications, Easy anti-cheat, and Battleye. Battleye is the problem. When you launch Fortnite, if the very first small splash screen has easy anti-cheat in the corner, then the game will work. If Battleye is active instead, you get booted. I want to create a new Q35-7x VM, but I don't have the NVME disk space. Really hesitant to blow out my current primary VM since I don't think I can reactivate some older software any longer.
  10. And I'm being blocked again. Tried physical NIC passthrough with no luck. Not sure what they are looking at... I also received this very helpful reply from support.
  11. Looks like switching to e1000 NIC was the ticket. Kind of annoying, but whatever. I still need to back out a couple XML changes to be sure.
  12. I've been playing Fornite for years on my Windows 10 VM without ever having issues. A couple days ago I started being blocked with the message below. The only recent change I made on the VM was to install DaVinci Resolve to do some video editing. I uninstalled all that but still being blocked. I have not tried creating a new VM yet (low on NVME disk space). Is anyone aware of a workaround for this issue?
  13. Please keep it in CA. I plan to continue using it until the the iOS app is no longer supported, or if I need support for a new camera at some point.
  14. Looks like the plug has been officially pulled. Overall the impact isn't too bad for me. I run WireGuard and openvpn on an edgerouter. Guess I'll just ride that ship for now. I wonder if they'll yank app support?
  15. So login is still working, thought maybe we dodged the bullet, then I found the update below. Frustrating... nice while it lasted. Please keep the docker available since it will still work fine with local/vpn login.
  16. I updated from 3.10.10 to 3.10.11 a couple days ago as well, no issues.
  17. My upgrade to 6.8.1 was successful. Idle CPU use seems to have returned to normal. RX570 passthrough performance seems to be back to normal, I had been assuming performance issues were due to a driver update.
  18. I've noticed a few items failing to pull metadata automatically. I didn't make a connection to unRAID, but in hindsight it did start around 6.8.0 upgrade as well as other minor SMB issues. I haven't upgraded to 6.8.1 yet.
  19. I started having this issue on 6.8.0 with a Ubuntu VM with multiple cifs mounts. Adding "vers=v1.0" seems to have cleared it up. Not the preferred solution though...
  20. Okay I'm happy to report that i440fx/seabios is working for 11.3+. I swear I tested the installer ISOs, but I must have only tested i440fx with an existing install, not realizing GPT partition scheme is not compatible. Converting to MBR looks possible, but risky. My FreeBSD VM runs headless, so I'm not sure what the i440fx passthrough implications are. Official patches are on the way for Q35/OVFM.
  21. More progress on the FreeBSD side. Patch attached in bug report. https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=241774#c35 It really sounds like i440fx should be working on unRAID...
  22. After the reboot, you have to configure DHCP uname -a FreeBSD cylon12 12.1-RELEASE FreeBSD 12.1-RELEASE GENERIC amd64
  23. I believe the alias was added when I added a second e1000 interface for troubleshooting. It should work. At this point all I can recommend is start over and apply first patch and kernel without Netmap. <interface type='bridge'> <mac address='52:54:00:70:6f:f4'/> <source bridge='br0'/> <target dev='vnet1'/> <model type='virtio'/> <alias name='net0'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </interface> <interface type='bridge'> <mac address='52:54:00:7e:36:f5'/> <source bridge='br0'/> <target dev='vnet5'/> <model type='e1000'/> <alias name='net1'/> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </interface>