offroadguy56

Members
  • Posts

    17
  • Joined

  • Last visited

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

offroadguy56's Achievements

Newbie

Newbie (1/14)

1

Reputation

1

Community Answers

  1. Thanks JorgeB. Looks like the SATA passthrough to VM was the root of the problem. I'm not entirely sure how it managed to get that way. All I remember at the start from a week ago was plugging my GPU accelerator back in after changing it's cooler while also plugging in a 2nd m.2 NVME drive. The computer attempted to boot windows off that 2nd NVME as I had not wiped it. It tried several times before I caught on to it. After getting into unraid I noticed Disk1 was disabled so I restarted unraid multiple times and tried changing cables/SATA ports. When I disabled the array to fix Disk1 (just a simple stop array -> start array) I also simultaneously added a 2nd slot to my cache pool which changed it from xfs to btrfs which disabled my working cache drive (the 1st nvme). I don't believe loosing the cache pool was the cause or symptom as Disk1 was disabled before I touched the cachepool. But I could be remembering wrong because libvirt.img was in a share that was solely stored on the cache drive. So the SATA passthrough issue could have happened when I added that 2nd slot and drive to the cache pool which caused the file system to change causing the cache pool to become unreadable. Thanks again, the community here is great.
  2. Disk1 is down again. My docker img was corrupted at some point so I went to go fix that. I also had 2 images on the array, docker.img and docker-xfs.img. I deleted both then started the docker service with these settings. Do we know if this has caused Disk1 to go offline. This time it says it's enabled but unmountable: wrong or no file system. EDIT: I found a previous post referencing xfs_repair. I was able to execute the command and disk1 appears to be back and operational. waffle-diagnostics-20230307-1507 - removed corupted docker.img_then made new docker img as btrfs.zip
  3. Looks like the array is back online. SMB shares are working again. Docker and VM are currently disabled. I can work on my own to get those back. Can I remove these historical disks without further issue?
  4. I though the read check that it offered me would fix the disk being disabled. It did not. So I performed the start array with out disk and add it back trick. The array is performing a parity sync now. Parity drive is enabled now. I'll see how it goes, last time it did it's parity check but then the drive was disabled again, but that shouldn't happen now with the VM issue removed. I'll past back here with results sometime tomorrow after some sleep.
  5. Array started, libvirt deleted (i believe). Here is the most recent diagnostics. Here is most recent screenshot of Main tab. For when I attempt to fix the array. And my normal SMB shares have shown themselves on Disk1 again. No more linux file system. waffle-diagnostics-20230305-1047 - After libvirt deletion.zip
  6. Looks like the file was properly modified by the webui. But unraid failed to properly shut down the VM service. After a restart I can now modify the Libvirt storage location path. Do I need the array running to see the option to delete libvirt.img? If the SATA controller passthrough was the culprit, with the VM manager disabled in theory I should be able to start the array and repair it without issue correct? Even with out removing libvirt.img? There is also one more thing I want to point out. Currently on my Disk1 if I look at the contents of the disk I do not have my usual SMB shares. Instead I see a linux file system. If I navigate to /mnt/ I can see Disk2,Disk3,Disk4, etc but no Disk1. Just want to put this info out there before any more rebuilds or parity checks are performed.
  7. I have tried to disable the VM manager. It still says running in the top right. But the VMs tab is gone and enable VMs is set to 'no'. Unraid did hang on the loading icon for a few minutes. I refreshed the page to regain control of the webui. How should I go about removing the libvirt.img file? I assume I would see a button next to the path location in the settings page. This is the most recent line in the log:
  8. My VMs list is currently empty. I should have a windows 10 VM there. Any suggestions to accomplish what you recommended above? I feel like I've done enough potential damage. I'd like to play is slow and safe and see what the community suggests first. I assume I could fix this problem by disabling the VM manager?
  9. Unraid 6.11.5 I have some weird happening with my array. I will try and describe the series of events the best I can. TL;DR I tried to bring my disk1 online. It and parity disk took turns being offline. After 3 or 4 parity syncs and disk rebuilds. Parity disk will not go online after parity sync. Unraid has multiple notices saying Disk1 Disk2 can't be written to. Disk1 has read errors. Parity disk is disabled. Before I began my upgrade I have 7 disks total with 1 being parity. I have 1 cache drive NVME. I had planned to install a 2nd cache drive and assign it to the same pool. I install the 2nd cache drive. On bootup it asks me to assign it BTRFS. I click yes and that brings my 1st cache drive offline because it was formatted XFS. At the same time my Disk1 showed it was offline. I restarted Unraid multiple times and tried different cables and SATA ports. I though the disk being offline meant that it was not recognized by the OS/BIOS. I learn that was not the case and that the disk had to be rebuilt by the parity data. I take the array offline remove the disk, start the array, stop the array, add the disk back. Rebuild begins. After rebuild unraid says all disks + parity are online. I remove the 2nd cache slot and set the cache pool back to XFS and assign my cache drive back to the pool. I move the data off the cache drive by invoking mover and setting all of my "prefer cache" disks to "yes cache". All the data moves successfully. I also as a precaution copy the appdata folder contents to my main PC via SMB. I notice that my VM list and Docker list are empty. I restart unraid. I should mention at this point and before during the restarts Unraid was not able to properly shut down. It would either hang on trying to stop the array, and do absolutely nothing for 30 minutes or more. Or it would spit out on the local console that it had IO errors. I didn't think anything of it. 3 times now upon boot up my computer would not recognize any bootable devices including the unraid USB except for the 2nd cache drive which had a windows install. After a restart unraid would boot. After actually booting into unraid if I had rebuilt disk1, the parity disk would be offline and I would take the array down and back up to get it to parity sync. If I had previously done a parity sync before a restart disk1 would be offline and I would take the array offline and back online to rebuild it. Now after each restart and parity sync, the parity disk remains offline even before the restart and the "successful" parity sync. Also as of now there is no longer a parity sync button, it has been replaced with a read-check button. I'm sure some info has been left out and that certain things aren't very clear. Please ask me questions and point me in the correct direction to recover my array. My only hypothesis is that I somehow swapped my parity and disk1 positions. I have attached some diagnostic dumps and here are screenshots of my current webui. I also have a flash backup of unraid from 02-15-2023 I believe it is version 6.9.5 in that backup? Thanks in advance for all that yall do here. waffle-diagnostics-20230303-1230.zip waffle-diagnostics-20230303-1932 after parity sync.zip waffle-diagnostics-20230303-2004 after restart to normal OS mode.zip
  10. The card is 24gb. I didn't think it would be an issue on address size. I still don't really know how all that works. When I was searching for code 12 errors I didn't see any direct correlation to address size so I pushed those results aside. It didn't help that most results weren't for my use case anyway. Again, thanks a bunch.
  11. I ended up finding my solution: Corsair Commander Pro for USB control of a single PWM fan. Pass the USB device through to the VM. iCUE unfortunately would not detect my GPU temperature. Instead I used this software which detected both the fan and GPU temperature. https://github.com/Rem0o/FanControl.Releases Not the 100% free and purely software solution I was hoping for, but it works.
  12. Holy smokes, it's worked. I'll begin testing actual functions of the card, but for now Windows is no longer complaining and GPUZ/Afterburner have recognized it. Even Nvidia control panel is working. How did you end up finding that post? I didn't think a card from 2016 would have resizable bar. My GPUZ says it does. Thanks so much!
  13. I recently bought a used Tesla P40 for AI work with Stable Diffusion. I planned to run it in my Windows 10 Pro VM. However after installing the Telsa P40 data center drivers from Nvidia Windows will recognize the card but will display a Code 12 error saying it cannot find enough free resources. I am completely stumped at this point. The GPU works perfectly fine when running bare metal. But as soon as it's passed to a VM it's not functioning. I tried Linux Mint at one point, drivers installed successfully (as far as I could tell) but Mint would notify that it can't detect an appropriate GPU. Any help would be appreciated. If you have questions let me know. Here is my VM setup and the XML
  14. unRAID version 6.9.2 I am looking at getting a datacenter GPU for ai image training for cheap. Much cheaper than what is normally available as a normal graphics card. It has no fan or headers on the graphics card board. I plan to 3D print a duct and attach my own fan and connect it to my motherboard's case fan header. I have Dynamix Autofan and temperature plugins. There doesn't seem to be any option to reference a specific temperature, I just see mainboard or CPU. And Autofan just lists options for connected drives. Is there a plugin or some software I could run at unRAID's level to control a case fan determined by the temperature of a specific probe, like the GPU? The only other solutions I can think of is a.) external fan controller with some auto fan curve and a temp probe stuffed into the GPU. or b.) USB fan controller, like Corsair Commander, and pass the USB device to a windows VM. The performance of the external fan controllers I can find on amazon aren't great. Users complain of not being able to adjust the fan curve or that the curve is incorrect for their use case. And for my Windows VM, I am not able to currently pass through fan speed and temperature info for the software without that USB controller. And of course last resort is just power the fan at 100%. But this server is in a public space in the house, I have no server room near a LAN jack. So I've only been able to keep it around with how little noise it makes.
  15. Ok thanks for the explanation. I have a bit better understanding now and will play around with your suggestions in time.