• Content Count

  • Joined

  • Last visited

Community Reputation

32 Good

About Warrentheo

  • Rank
  • Birthday October 5


  • Gender
  • Location
  • Personal Text
    Currently running:

    unRaid 6.4.1 Pro License Since 02/02/2018

    Asus IX Hero with i7-7700K and 64GB RAM

    Windows 10 Gameing VM with GTX 1070 passthrough

    5HD's with m.2 Raid-0 Cache

Recent Profile Visitors

1543 profile views
  1. From his lsscsi.txt file: [0:0:0:0] disk JetFlash Transcend 16GB 1100 /dev/sda /dev/sg0 state=running queue_depth=1 scsi_level=7 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/0:0:0:0 [/sys/devices/pci0000:00/0000:00:01.2/0000:01:00.0/0000:02:08.0/0000:05:00.3/usb4/4-4/4-4:1.0/host0/target0:0:0/0:0:0:0] [6:0:0:0] disk ATA WDC WDS500G2B0A 90WD /dev/sdb /dev/sg1 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/6:0:0:0 [/sys/devices/pci0000:00/0000:00:01.2/0000:01:00.0/0000:02:0a.0/0000
  2. Ah sorry, I knew it defaulted to 2 Parity, but I thought you could select more from a drop down underneath the 2 like you can with the data drives... Haven't been able to shut my array down to double check... Maybe you would want to replace and rebuild one drive at a time then, that way you are not without parity protection... Just in case...
  3. I think you can just shutdown the pool, and add as many parity drives as you wish, if I remember correctly... Once the new parity drives are up and running, you can use the "New Config" tool in the tools pane to take drives out of the pool, and just check the "Parity drive is already up to date" to turn the pool back on... afterward, you can use the old drives as you see fit...
  4. Update, 10 days later... It appears that the data for every drive in the pool somehow got very corrupted... I am very miffed about this (around 11 TB possibly lost), but I want to be clear ahead of time, I am NOT expressing anger at UnRaid or Lime Technology, since I am sure about 95% of this is my fault somehow... My Main goals here are to figure out how this happened in the first place, and if possible find a way to keep it from happening to someone else... I realize there are several steps where I got punch happy and did things I even knew at the time I should not be doing... That said, tot
  5. Both gui, SSH and even local direct all show the same info, the data isn't in the root of the drives anymore, No files were moved in the main pool ( I just emptied the last few files off the cache on it ) Files were there before the hard reboot, and were all gone afterware Thank you for the replies, I am out of school, and will get home and try some of these suggestions and see where we are at...
  6. The cache was used primarily as a cache, and VM virtual disks, with some other minor knick knack files thrown in... 98% of the data on the server was under the main pool stored on magnetic media, only 1 share was set to "Cache Only"... I disabled VM machines, disabled Docker, deleted their associated image files, and had moved all other files off of the cache... it was reporting empty when I attempted to stop the pool, remove the cache drives with the "New Config" option (and only the cache drives were dropped) and perform a blkdiscard on the cache drives, I then attempted the reboot, which
  7. I have checked the individual /mnt/disk?/ locations, they also reflect the same 2 shares that are reflected under /mnt/user/ I am trying to find a way to restore the original files if possible... or possibly find where the files got lost to and resetup the pool and copy things back manually...
  8. That occured just before the issue and was intentional, but I don't think it is related to the actual issue... I had corrupted the cache pool about a week ago, and repaired the issue, however I ended up deciding the repair was less than ideal and decided to format are restore from backup the data I had on it... The format worked as intended and I was getting ready to perform the restore, I issued a reboot command, when the system locked up for almost 15 minutes (longest previous reboot was around 5 minutes) I was unable to shutdown with any method, so I was forced to hard shutd
  9. Hello, I need some help, I was forced to hard reboot my server (I run a windows gaming system on top of it as well). when the system returned, only 2 of the 25 or so file shares are now present... I have booted the array into maintenance mode, and did an xfs_repair on each drive in the array, but this didn't seem to help... I know a bit about linux, but I also know that I should ask for help before proceeding... The drives appear to still have the correct amount of space on them consumed, so I think the data is still there, I just can't seem to find a way to safely access it...
  10. I just got notified that the plugin is not known to the community apps plugin or the fix common problems plugin, and when I search for it now, it doesn't show up in the list of plugins that can be downloaded.... When I search for "NVIDIA" in the community apps plugin, it no longer shows up... Rebooting the server, and uninstalling the GPU statistics plugin changed nothing... Did this plugin just go unsupported? or is there a glitch in the community apps plugin? if it is now unsupported, I will be very sad to see it go... 😢 Still thank you to the author no matter whic
  11. It is a 2Tb pool, 80% full, and was originally in raid0, the only issue it had is being booted once with one drive missing... I am trying to see if it is possible to start it again in raid0, or is the array already corrupted?
  12. That is the correct, the question is is it possible to skip the conversion to the normal raid1 mode, and leave it in the raid0 mode...
  13. Hello, I recently moved locations, and during my move, on of the cache drives that I have in a raid0 btrfs pool came disconnected... Array got started with only one of the 2 drives... After troubleshooting, I discovered the issue, and stopped the array, on the main screen the drive had become "unassigned"... When I re-add the drive, UnRaid now warns me that the pool will be formatted when the array is started (I think it is trying to boot back in normal mirror raid mode, and then will make me switch it back to a raid0 array) Is there a way to just re-add the drive in the original r
  14. I don't think Windows even thinks in the terms of IOMMU groups, most likely you will need to temp boot off of a Linux Live USB/DVD image... The commands are: for d in /sys/kernel/iommu_groups/*/devices/*; do n=${d#*/iommu_groups/*}; n=${n%%/*} printf ' IOMMU Group %s ' "$n" lspci -nns "${d##*/}" done You have the setup I am looking into getting as well, so that info would be useful 😄
  15. I also have this bug, rc5 to rc6, VM's show fine until one gets launched (either from GUI or CMDL), then it goes blank... I only have one VM, and so I am willing to rebuild the libvirt partition for testing if needed... qw-diagnostics-20191118-1759.zip