axeman

Members
  • Posts

    538
  • Joined

  • Last visited

Everything posted by axeman

  1. did any of you guys have a permissions issue starting the VM? I kept getting a Permission Denied error trying to start the VM...
  2. do you remember what you did to get this fixed? my VMs are all topping out at about 70Mbit ... even though I have 400Mbit. If i boot Unraid with GUI, i can get about 360Mbit, which is what I used to get in my VMs (same hardware) on ESXi. Even creating a new VM doesn't seem to help.
  3. Happy quarantining everyone! I'm looking at migrating from my esxi 6.0 rig to kvm. At the moment UnRaid runs inside esxi as a vm. So in prep for this, I figured I run a few simple tests. I backed up my UnRaid Flash to a folder on a physical machine. I converted a couple of vms from within the unraid vm via the qemu command to convert from vmdk to raw. shutdown esxi, and booted directly to unraid on baremetal. ran a couple of configs (network, etc) started up the VMs...created a new vm, etc. All worked great. But i'm not ready to move over just yet. weekend ran out, and need to revert back. Simply shut the server down, took the flash over to my physical machine, deleted all the files (not format) - and copied from my backup. All good so far. Unraid boots up, and now shows the VMs that I was playing around with. QUESTIONS: 0. who is your daddy and what does he do?! (sorry ignore this one). 1. How does it know that? If I replaced all the files on the flash? 2. Also Windows machine was asking for Activation, because of the obvious hardware changes - when I eventually do it, will it activate, or will I basically give up my windows license? 3. The linux vms that I have are running things like ConnectWise, PiHole and OpenVPN - all of which I think are better to just start over and migrate configuration as opposed to converting the disks.... thoughts? 4. when creating a linux vm - i saw the option to mount an unraid share... i tried that, but couldn't find it mounted anywhere. I could still go to network and navigate to it, but that's not exactly what I was looking for. I thought it'd show up as a folder or something... where does it end up? 5. finally - i can probably look this up - but the linux mint VM kept complaining about the display driver... but I couldn't see how to update it... how do i get it to use the libvirt drivers? Sorry - lots of nooby quesitons here. My processor is an ancient Intel i7 960 ... yes, it's 10 yrs old, but still works great with multiple vms running. I just want to move off esxi since they stopped supporting it after the last update. Plus I figure remove one extra layer of abstraction can't hurt.
  4. Mods - you can delete this. I tried, but can't seem to.
  5. I put the disk to be replaced back in... and powered everything on again... This time, i got valid configuration, and prompt to enter encryption key.
  6. I had a failed drive that I was planning on pulling out and swapping in a new drive. I stopped the array, and powered down my drive chasis. This was probably stupid. As I did that, Unraid started throwing a bunch of errors about drives going missing. So I did a clean shutdown. After that, I powered the chasis back up, then powered UnRaid on. Upon first boot, Unraid is asking me for a NEW encryption key. At the bottom, it says Stale Config. I want to proceed carefully. The drive that was pulled is still available. What's the safest way to proceed?
  7. A bit of an update.... I put the Cache drive and the former Disk 10 into my other array, and ran some pre-clears. Believe it or not the Cache drive (Samsung 1TB) ran through and passed. Disk 10 faltered... and there's something definitely wrong with it it's going, but VERY slowly. reads are at about 18-20MB/s on the same slots where the Samsung was cruising at 100+MB/s. Meanwhile, back on the main array, Disk 5 (WDEARS) has tons of read errors. I have the two pre-clearing (they are the largest in the array, so will end up in parity), so that won't help. For the moment, I'm reading data off the disk and moving to another array disk. Errors are at 39K and rising. I'm just hoping that I can move most of the data off before the drive dies. Perhaps remove it from the array (shrink), wait for the disks to be pre-cleared and re-expand. I know parity doesn't need pre-clear, the reason i'm running these here is because they are WD shucked, and want to make sure they are good in case I need Warranty support, etc.
  8. Thanks again - my original rely to you didn't get posted. I have two drives enroute. A parity scan (non correcting) is running. I'm going to pull these dead/dying drives out and put them in my unraid bench setup to keep diagnosing. Don't want to risk data loss on the main array. I appreciate your assistance throughout this whole ordeal.
  9. oh weird... well, best i can get is the short test. the long test doesn't get past 10%... I'm unbalancing some data around to get the data off Cache ... and boy are you right with that one too.. the HD light is solid on, so definitely HW failure. just glad this is an Unraid array, otherwise, I would not be so zen about it all. TOSHIBA_DT01ACA300_X3G7E59KS-20190220-0941.txt
  10. Well that's certainly not what I expected when I created this thread! Okay. So I'm going to run a extended smart test on Disk 10.. .that's the only spare I have ATM. It's got a bunch of reallocated sectors though - so might not be the best candidate. I guess the question would be what's in a worse state, Disk 10 or Disk 5 (while I get a replacement ordered, delivered and installeD). the Cache pool is the ONLY part of the server that I don't backup - since I always assumed it'll be data in flight.
  11. AND we have smart errors on Disk 5 as you predicted. rigel-smart-20190219-1500 DISK 5.zip
  12. Thank you! I ran the scrub, and results look awful. scrub status for 8844a357-4626-49a0-a8f2-797ad43999be scrub started at Tue Feb 19 13:39:12 2019 and was aborted after 00:25:00 total bytes scrubbed: 67.70GiB with 8854316 errors error details: read=8854312 super=3 verify=1 corrected errors: 78, uncorrectable errors: 8854235, unverified errors: 0 I didn't abort it, i'm guessing self aborted? I'm going to run the extended tst on D5. I appreciate your input!
  13. Thanks for taking the time to go through it. Strange - Cache Pool had a device that got the click of death about 6 months ago. I removed it. I will run a scrub on that. I know one of the destination disks have been full (so my mover isn't moving). Could the errors here be because of that? I'll read through your link and see what it turns up. Unraid has not kicked Disk 5 out of the array yet - should I run a parity check? or test the recently removed Disk10 and use that as a replacement for disk5?
  14. And another disk shows read errors ... but is not disabled, so that's good. Attached is new diagnostics. So clearly, I have either a cabling, power or controller issue. It'll be fun trying to narrow down where the failure is. Incidentally, I haven't had any errors until going to 6.6.6 - maybe it's more diligent in finding these? rigel-diagnostics-20190219-1120.zip
  15. Disk 4 had some VM HD's on it - I botted up the VM and scanned, everything seemd OK. I'm currently rebuilding Disk10 on the newly cleared drive. I'll probably put the original Disk10 into the Cachepool Thanks for talking me through it.
  16. Thanks.. should I expect some sort of corruption, since we're emulating two disks, or did dual parity prevent this from going south. I agree that it's probably related to controller/cabling/power. Disks seem OK otherwise. Should I run a parity check after the rebuild, or start rebuild of disk10 and then run a parity check?
  17. it's been a while.. is there anything in particular I should post? the diagnostics download doesn't seem to complete. smartctl on the two failed drives? syslog? Thanks again for your help
  18. Okay Gents... a doozy for you: I had disk 10 show errors for a while, and it eventually got disabled. I got a replacement, and pre-cleared it. As it was clearing, I got lots of umda crc errors. I feel like this has happened whenever I preclear a disk, i get random umda crc errors. Of course it was probably a power or controller issue (as of the cable reseating, etc I have not had a single umda-crc error). As the pre-clear was about to finish- disk 4 in the array got redballed. No biggie, the pre-clear finished, I resated all cables, shutdown array, unassigned disk 4, and reassigned it. The plan was to assign the newly cleared disk to Disk 10. As Disk 4 was getting rebuilt (18%), I get a notice that Disk 10 has read errors and is disabled. Thankfully I have a dual parity setup. I think I'll be OK, because of this (unless another disk gets disabled). Question is - what do I do? Shut down, stop disk 4's rebuild and have unraid rebuild disk 4 AND disk 10 (with the newly cleared drive)? Or wait for disk 4's rebuild to finish? Do I allow that parity check to finish after the rebuild, or immediately stop and unassign, and reassign disk 10 with the newly cleared rive? I'm at a precarious position, obviously - want the safest course of action to mitigate the situation... FWIW, i'm on 6.6.6
  19. Thank you! Safe to do while the parity check is going, I assume?
  20. I tried ... but it timed out. As a matter of fact, all of UnRaid seemed to be a bit wonky. The UI wasn't accurately showing spun up/down drives etc. The physical clicking thankfully came from an older drive that was part of the cache pool. I had to do a reboot... while it looked like it went down OK, when it came back up, I got an unclean shutdown warning and let the parity check get started. There was something definitely not right, because I had a parity check 2 days ago, get the UI was showing last check at 84 days. I'm on 6.4.1, and havent upgraded yet, but after all this blows over ,will. NOW - the real question... since it's one of my cache pool drives that died, how do i start it back up safely? It seems like the cache pool is inaccesible.
  21. So one of my drives started clicking this morning. It's an actively used drive - the thing is UnRaid has not detected it as failed yet. All I can get from the log is this: May 28 06:03:10 Tower kernel: blk_partition_remap: fail for partition 1 May 28 06:03:10 Tower kernel: btrfs_dev_stat_print_on_error: 55 callbacks suppressed May 28 06:03:10 Tower kernel: BTRFS error (device dm-15): bdev /dev/mapper/sdu1 errs: wr 18388755, rd 18966224, flush 1256, corrupt 0, gen 0 May 28 06:03:10 Tower kernel: blk_partition_remap: fail for partition 1 May 28 06:03:10 Tower kernel: BTRFS error (device dm-15): bdev /dev/mapper/sdu1 errs: wr 18388756, rd 18966224, flush 1256, corrupt 0, gen 0 May 28 06:03:34 Tower kernel: usb 1-1: reset high-speed USB device number 2 using ehci-pci I have a dual parity setup, and would like to assign Parity 2 drive to the failed slot... what's the best way to proceed? Thanks! Sorry to disrupt the somber weekend.
  22. Thanks - the parity sync is still going, found 792 corrections - probably the highest I've ever seen. Then again, haven't really lost power in the middle of a known write-heavy operation. Love how UnRaid keeps on truckin! I've lived in this house over 5 years now... only lost power 3 times ... Sandy, another major unnamed storm, and of course yesterday. Had to be in the middle of this. Oh well.