timekiller

Members
  • Posts

    78
  • Joined

  • Last visited

Posts posted by timekiller

  1. Sorry for the long post, but I an trying to get all the information into 1 post so there isn't a lot of back and forth.

     

    I am trying to upgrade my Unraid server with a new CPU/MOBO and I'm having a hell of a time. My existing hardware is a Core i7-6700k. I have a Ryzen 5800x CPU and I've been trying to find a motherboard to match with it that will support my existing hardware.

     

    Specifically I have (2) IO Crest 16 Port SATA cards, and (2) 1TB m.2 NVME drives for cache. My existing motherboard (MSI Z170A), which is like 5-6 years old at this point is able handle all this hardware without issue.

     

    I have now bought 2 motherboards (first board details and issues are in this thread) 

     

    I returned that board and now I have a Gigabyte X570 UD which almost fits all my hardware requirements. I'm ok on PCI slots, but there is only 1 m.2 slot. Not a HUGE deal, I can buy a new 2TB m.2 drive to replace my (2) 1TB drives.

     

    The problem is that now this board will not show any drives from the SATA cards. I have tried several configurations, including removing the m.2 drive and only installing 1 SATA card. No matter what I do, I can not see any drives attached to any SATA controller on this board.

     

    The UEFI screen is very stripper and does not have a section to see installed peripherals, so I can't even confirm that the card is recognized by the system. When the system boots, I do see a ton of errors related to ATA, so I'm not sure if my issue is at the bios level, or if there is something up with the UNRAID configs.

     

    I did pull a diagnostic dump and attached it to this post.

     

    Any help is appreciated.

     

     

    storage-diagnostics-20211009-0128.zip

  2. Correction. The slots are as follows:

     

    (1) PCIE x16 slot (PCIEX16)

    (1) PCIE x16 slot (PCIEX2)

    (1) PCIE x16 slot (PCIEX1)

    (1) PCIE x1 slot (PCIEX1)

     

    So the second PCIE x16 slot is actually only an x1, this might explain why on boot Unraid is showing a bunch of "limiting drive to 3Gbps" lines on boot.

     

    Not sure if the mismatch of x2 and x1 for to two identical cards is causing the issue, but now that I know the slot is only x1, I don't want to use this board since half my drives would be bandwidth limited.

     

    Going to yank the board and return it for something better. Guessing I'll have to go to a X570 to get all the PCI lanes I need.

     

    Thanks everyone!

  3. I was able to swap the board back in, but added no m.2 drives. Still have the same problem. Only 1 SATA card is recognized. It shoudln't matter, but I have (2) IO Crest 16 Port SATA III PCIe 2.0 x2.

     

    The motherboard has (3) 16x slots, but only the top one is true x16. I have a GTX 1050 int hat slot to help with video transcoding. The other 2 slots look like x4 based on the pins I am able to see. I'm consulting the manual to see if it tells me anything.

  4. I was attempting to upgrade my unraid server, but unraid will only show me drives from one of my sata cards at a time. The cards are fine, and if I switch them around, the drives I see changes (I always see the "second" card). If I remove the second card, I can see the first card. I figure this must be a bios setting issue, but can't figure it out.

     

    If I go into the bios, I can see both PCIe sata cards (see picture). I have attached my diagnostics for reference.

     

    Old Hardware:

    Intel Core i7-6700K @4.00Ghxz

    MSI Z170A Gaming M5 Version 1.0

    32GB DDR4

    2x 1TB NVME cache pool

    2 parity drives, 19 data drives

     

    New Hardware (NOT WORKING):

    AMD Ryzen 5800X

    B550 Aorus Elite V2

    32GB DDR4

    2x 1TB NVME cache pool

    2 parity drives, 19 data drives

     

    Thank you!

    20211005_165254_HDR.jpg

    storage-diagnostics-20211005-2036.zip

  5. 2 hours ago, UhClem said:

    I've looked into "staggered spinup" for a DIY DAS. The key search term you want to research is "disk PUIS" ... Power Up In Standby. It looks a little tricky, but quite doable. (That isn't a solution for me because my drives are connected via a SAS expander.)

     

    [I don't use Unraid.]

     

    Yup, I know "PUIS" because that's how Highpoint refers to it. I said "staggered spin up" because that a term that I figure more people would recognize. I've been searching, but either this is not a common feature, or it's just not documented well, I can't find much...

  6. My unraid server has 21 drives and the power supply can't handle spinning up all the drives at once. I'm currently using a Highpoint 750 (40 port) sata card but it appears to be having issues and won't be supported in the next release of unraid.

     

    Looking for a card that supports staggered spin up and ideally up to 30 drives. If I have to buy 2 16 ports cards, that would work too, but the key is that it MUST support staggered spin up.

     

    Before it's mentioned - no I don't let my drives spin down when idle, so yes the only time they all spin up at once is at boot. Yes I can upgrade my power supply, but when I hit 30 drives, I'm not sure a single power supply will have enough juice.

  7. 2 minutes ago, trurl said:

    I doubt there is anything wrong with that disk, SMART in diagnostics looks OK. Same for disk7 and 9. I guess you could run an extended SMART test on them, but since you are having multiple disk problems I think it more likely that the disks are not the cause.

    Hmm, I suppose I could try swapping in my old controllers to test. If it's the Highpoint that would be unfortunate since I bought it on ebay a few months ago. Though at least that would give me an excuse to get something that will be supported in the next version of unraid.

     

    Thanks @trurl

  8. Quick update:

    I powered down and reseated cables. Powered up and immediately disks 7 and 8 had problems again. Same as before, disk8 disabled, disk7 read errors.

     

    I powered down again and connected disk7 and disk8 to different ports on the sata card. Power up and disk8 is immediately offline gain, but disk7 looks ok.

     

    Data rebuild can now continue as there is enough parity information to rebuild. Fingers crossed that disk7 stays ok - only time will tell.

     

    If disk9 completes the data rebuild (in about 3 days) then I can power down, swap out disk8 and see what's up (probably have to RMA it).

  9. thanks @trurl I was dealing with some other (non unraid) stuff. Everything you said makes total sense. Going to shutdown and check cabling now. Fortunately I am backed up. I sync everything to google cloud regularly, and the replaced drives have not been wiped yet, so worst case I can swap them in and create a fresh array and I shouldn't lose anything. Also, yes I can see everything you were talking about on my main screen:

    unraid.thumb.png.efa8f34cd0ee1a724e7f247d9541df57.png

  10. I am in the process of replacing drives. I removed a 4TB drive and replaced with a 12TB drive. I'm 32% into a data rebuild and now I have a unraid reporting that a drive is disabled, contents emulated. My logs show a ton of messages like:

    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979624
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979632
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979640
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979648
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979656
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979664
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979672
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979680
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979688
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979696
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979704
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979712
    Jan  6 10:31:30 Storage kernel: md: disk7 read error, sector=5384979720

    I know I need to address this, but I'm nervous about doing anything while the data rebuild is running. Fortunately I do have 2 parity drives, so I should not lose any data but I'll feel a lot better when the data rebuild is complete (in 2 days!).

     

    What has me concerned right now is that I can't write to any share under /mnt/user. I can read from the array and if I write to a specific disk at /mnt/disk#/<share> I can see the new content when I access the share through /mnt/user/<share>.

     

    What I need to know is: Is it safe to let the data rebuild continue? Am I better off shutting down and seeing what is up with disk7? If I shutdown and it turns out disk7's sata cable is lose or something else like that then would I create more issues with the disk suddenly coming back?

     

    The nightmare scenario I'm imagining is if a new file is written while disk7 is offline, then I shut down and get disk7 back online, then when I boot back up the parity drives will have the wrong calculations based on disk7 and the data rebuild would be corrupted. Is this a valid concern?

     

    Diagnostics attached

    storage-diagnostics-20210106-1339.zip

  11. 51 minutes ago, JorgeB said:

    also have you tried without encryption?

    Are you asking if I've backed up 122TB of data, disabled encryption, and copied the data back to test disk performance? No. I have not done that.

     

    Is there any way to test the read/write speeds of individual disks without disabling the array and testing each disk? I had thought about writing a test file to each `/mnt/disk#` but realized with the array started, I'm still getting hit with parity calculations, so even if I'm writing to 1 disk, I'm reading from all of them and writing to the 2 parity drives.

     

    I'm wondering if anyone else has seen a performance hit when dealing with this many drives? Thinking about biting the bullet and ditching the 10 4TB drives in favor of 4 12TB drives. I'd get a capacity bump and get rid of 6 spinning drives which should help parity performance as well as heat and power draw. 

  12. I have an issue where Unraid is becoming unusable with extremely high load, caused by shfs spinning out of control. This seems to be due to radarr doing file operations, but I'm not 100% sure. What I'm seeing is the load on the server starts climbing (getting up to 150 at times), and all other actions that require disk access just hang. I can't even `ls` a directory. I've done all I can think of to handle it, but nothing has worked.

     

    Things I've tried:

    • Convert cache from Raid1 to Raid0 - I have (2) 1TB nvme drives for cache. My Plex appdata folder takes up 700GB alone, so very little space was left for the data cache. I  converted to Raid0 so I have 2TB of cache, but that didn't help
    • Move NZBGet, Sonarr, and Radarr off the unraid server. When content is downloaded is when it's the worst. Especially if NZBGet has to repair a file. I thought by moving these service to another system, I could offload that work so Unraid wold only have to do be available to receive the copied file (over an NFS share)

     

    These have not solved the problem. I still regularly see the load spike up. When that happens, I con't do anything with the server until the load drops. No file access, no Plex, nothing. It's truly infuriating.

     

    I do have a lot of drives (21 including 2 parity drives), but I feel like this has gotten way worse recently. I thought maybe I have a failing drive that is causing the parity functions to hang, but I am seeing no SMART errors that I would expect to see.

     

    I'm at my wits end and need help!

     

    Unraid 6.8.3

    21 spinning drives (19 data + 2 parity) 136TB capacity

    2 1TB nvme cache drives in RAID0

    storage-diagnostics-20201207-0935.zip

  13. Understood, This is mainly a media server, so I can have files spread out as typically only 1-4 files will be getting read at a time, and which ones will be pretty random. Besides, call it OCD, but I can't stand to have all my drives almost full and another drive with nothing on it :)

  14. Hello!

    I have an array of 13 disks (11 data + 2 parity). I previously had a drive I dedicated to my security cameras (disk 11). I had all shares except 1 "Exclude" that disk so only camera recording would go there. I have since removed the docker and removed disk 11 from all share "Excludes". However, no data is getting written to the disk. It's been weeks,  and it definitely should have been used by now.

     

    Screenshot and diagnostics attached!

     

    Any help is appreciated!

    Selection_199.png

    storage-diagnostics-20191029-1409.zip