Jump to content

DesertCookie

Members
  • Posts

    68
  • Joined

  • Last visited

Everything posted by DesertCookie

  1. As far as I can tell Unraid can write to the thumb drive. What permissions should all of the folders have by default? Maybe some folders work while others don't and I keep testing the wrong ones.
  2. I'll try to remove the drive. It's an M.2 NVMe SSD. This can't be the only reason though as this has been happening far longer than the drive is installed.
  3. Here it is. Thanks for pointing that option out to me. stower20-diagnostics-20210924-1640.zip
  4. I've been having the issue of Unraid starting a parity sync every time I restart my server - which I had to do multiple times lately because my USB acted up. I've read through this post on the topic but it didn't help me out very much. Could someone point me to the place in my log that tells me what exactly is going wrong so I can learn how to read the log and troubleshoot on my own in the future, please? My config: TR 1900X MSI X99 Taichi 2x 16GB DDR4-3200 ECC 1TB WD Blue NVMe SSD (cache) 256GB WD Black NVMe SSD (cache 2) 2x 12TB WD White (parity, array) 2x 4TB WD Se (array, array) 1TB Seagate (unassigned devices - location of games for VM) 32GB Samsung flash drive stower20-syslog-20210924-1339.zip
  5. I've retested today - after having made no changes, not even shutting down the computers since the last test. Now I get the following results: 1GbE: 412 Mb/s (spread of 277 to 859 Mb/s) 2.5GbE: 508 Mb/s (spread of 432 to 567 Mb/s) While this is better it still doesn't represent what I would expect from this set up. Edit: A little while later I found that 1GbE is mroe reliable. Whether it's problem of the ethernet cards or one of Unraid with the ethernet cards, I don't know. I'll be adding the name of the cards to the original post as soon as I stop my server for the next time.
  6. The 2.5GbE link is direct, as I wrote. I only have a phone via WiFi 4 available as second client (or a VM in the server but that's obviously useless here). I didn't change MTU as I didn't expect that to be relevant for only 2.5GbE. Could that cause an improvement?
  7. Using SpaceInvader One's video as a guide I got the following results: 1GbE: 74 Mb/s (spread of 38 to 120 Mb/s) 2.5GbE: 126 Mb/s (spread of 70 to 191 Mb/s) iperf3_results.txt
  8. Edit: What ultimately solved my problem was reinstalling Windows - now I'm back on 1GbE but at least get around 100-110MB/s. The Windows installation was about a year old and only had seen new RAM and some new HDD and SSD storage. Let's hope the upgrade to 10GbE later this week goes well. I recently installed a 1TB WD Blue M.2 NVMe SSD as cache, as well as a DeLock 2.5GbE network card and directly connected my PC (via its onboard 2.5GbE port to my server). Read and write speeds are down compared to before the update. Here are some metrics: 20-40MB/s write and 20-80MB/s read (the same as via 1GbE but more stable) read and write speed are pretty much the same for array and cache (verified using the web UI) HDDs are capable of 80-170MB/s r/w (according to DiskSpeed docker) - averages around 120MB/s cache SDD is cabable of 1GB/s-1.1GB/s r/w (according to DiskSpeed docker) 2.5GbE is direct connect Cat.7 (1m), 1GbE is via two switches and a router (total of 25m of Cat.6) <1ms latency via 2.5GbE instead of 2-3ms latency via 1GbE previous no-name 240GB M.2 SATA SSD was a lot faster Overall, the web UI feels a lot snappiere since accessing it via 2.5GbE - before traffic needed to go through a switch, to my router, and back to a switch - now it's direct. Windows 10 shows me 2.5GbE speeds. The M.2 SSD is connected via PCIe 3.0 x4. I really don't know where the bottleneck is here. Internally (e.g., copying from array to cache) is faster but not as fast as I expect from this hardware. I'd be grateful if someone else could take a look at this. I'm happy to provide more information - I just don't know what more at the moment. My system: ASRock X399 Taichi Threadripper 1900X 2x16GB DDR4-3600 (@3200 1.25V) - survived a memBench run 1x WD Blue 1TB as cache 2x WD Se 4TB as main array drives (no parity atm) The system I connect from is a Ryzen 9 system with a B550 Mainboard with onboard 2.5GbE - plenty of power on that end. syslog.txt
  9. Just for future passers-by: I had the drive I had issues with mounted via the Unassigned Devices plugin and had forgotten to unmount it.
  10. Alright, the repairs runs correctly. I don't know if it actually did anything to releviate the drive errors that seem to move with the data from disk to disk.
  11. I used the name displayed in the web ui of "sdh". I used "/dev/sdh". I see how I might have messed up...
  12. I've pulled both out of the array for the moment; I've ran the extended diagnostics - the original drive with known bad sectors wouldn't even run it. I got errors on a completely healthy drive that went into error mode too. I'll try swapping the HBA card asap! It's an Adaptec ASR-71605 and it's not actively cooled. I suspect that might have caused issues as those are known to run hot. I'm running xfs_repair -vn on these 4TB drives. Is it normal for it to take multiple hours and only display dots for an extended period of time? There have been 2M reads on the drive so far.
  13. I will try a reseat when I have physical access to the hardware again. I still have two free x16 slots I can try. For now, after a restart, it picks them all up again. I have found the drive that threw up errors this time to also have some SMART alerts. I have appended the SMART report (first is said drive, second the other drive I already knew to have some issues). stower20-smart-20210203-2040.zip stower20-smart-20210203-2058.zip
  14. Looking at how I had a brand new flash drive fail after just a month I'm worried I'll be in a similar situation too at some point. With all the issues I'm having I'm wondering why I didn't stay with Windows Storage Spaces.
  15. After a recent monthly parity check, one of my drives went into error state. I pulled the data off the drive and removed the drive. Now, running the parity check I get a lot farther in the parity check but now a different disk is throwing 79k errors and went into error state. The parity check paused. The drive originally throwing errors wasn't the most healthy with about 70 bad sectors. It had amassed these bad sectors one year, and three years ago and hasn't gotten any new bad sectors since then. They all were corrected. The second drive's diagnostics are appended. I've had issues with a faulty flash drive recently and had rather a lot of troubles with Unraid. In the past month, I had to redo my drive configuration a bunch of times and thus ran more than half a dozen parity checks - these two being the only one's to throw errors. Edit: Fiveteen minutes after the disk went into error state all other drives connected to that HBA card are gone. I guess it's an issue of my Adaptec ASR-71605 then? Edit: Got an agent notification that the parity sync just aborted without input. It found 21,341,196 errors. What is the recommended course of action here? Edit: I zeroed the two drives throwing errors. One returned to normal operation according to SMART, the other got better. I'll use them for unimportant data. stower20-diagnostics-20210203-1852.zip stower20-smart-20210203-1851.zip
  16. I'm getting an error when accessing the web UI: Error 500 Internal Server Error | nginx . I can still SHH into it. My VMs are still running and accessible. Most Docker containers are running but my reverse proxy container is non-reactive and cannot be rebooted. I've tried shutting down the VMs with sudo virsh shutdown <name> - nothing happens and I abort the command with CTRL+C after a long time; I've tried shutting down all Docker containers with sudo docker stop $(docker ps -q) - my Nginx Proxy Manager container won't shut down (the rest does); then I tried stopping the array as per wiki - there's always active streams that I can't forcefully stop and thus cannot unmount the array. Finally, sudo reboot -f sudo powerdown -f do not work with no console activity like for the VM-shutdown. Apart from a hard-reset (which I cannot do as I'm not at home), what can I do? The last thing the server did was an automated monthly parity check which failed with error -4 according to the notification email. I tried collecting diagnostics but the command hangs like the VM-shutdown. This is a two-month old server and I've nothing but bad experiences with it: There has been no more than a week between severe problems that caused everything to spiral out of control. The one month of trial-period went without any hiccups which is what annoys me most. My system: Unraid 6.8.3 ASRock X399 Taichi, Threadripper 1900X, 2x16GB DDR4-3600 (@2400 as I suspected TR first-gen might not like the 3200MHz I had it before - was one of my first troubleshooting steps), GTX 1650 (for gaming VM) 4x 4TB WD Se, 1x 1TB Seagate Desktop HDD, 1x USB HDD (doing data-recovery on it), 240GB WD Green (SATA M.2, cache), Samsung 32GB flash drive Also posted on Reddit.
  17. Thank you. I'll order a new flash drive then. Strange, since it was brand new (Samsung BAR Plus 32GB).
  18. My system loses network access but seems to stay operational. Only a hard reset will get it back to be visible in my LAN - at least for a couple of hours. Strangely, the VMs still appear as online to my router, though it doesn't know what IP adress to give them or how they are connected (doesn't show LAN_1GB like for my other devices). Of course, a hard reset it somewhat undesirable. I'm not at home to troubleshoot myself. I asked him to log in and export the log and send it. This is what my homemate sent me instead: He said stuff like this was scrolling by. This means the VM visible in my LAN cannot be on because it uses this GPU. I asked him to reset the system. Now it says kernel panic -not syncing VFS: unable to mount root fs on unknown block(0,0) I've read about RAM being an issue. Also the USB breaking. Sometimes the USB just needed to be redone. Any other insights before I make the long way home to repair this one-month old system? Overall, I really like Unraid but have had regular issues with it. I'm a person that regularly gets broken hardware from reputable sellers or has stuff break - I don't know if it's me but computer issues and problems seem to gravitate towards me. Might be important: Restarted the server recently from the UI. Upon startup all array drives were in unassigned devices and Unraid didn't recognize them (classified them as new device when adding them). I added the drives back in like they were before and Unraid started rebuilding the parity disk. No data lost but weird. A few days later I added a second stick of RAM. Upon start the cache drive was not assigned. I added it and everything seemed to be fine. My system: Unraid 6.8.3 ASRock X399 Taichi, TR 1900X, 2x16GB DDR4-3200, GTX 1650, Adaptec ASR-71605, 4x4 TB HDD, 240GB SSD
×
×
  • Create New...