Jump to content

FrozenGamer

Members
  • Posts

    369
  • Joined

  • Last visited

Everything posted by FrozenGamer

  1. See attached diagnostics - i saw a PIO reference in here which i remember to mean shit disk transfer from windows years ago tower-diagnostics-20190120-1020.ziprs ago. But only that reference once. Jan 16 16:49:16 Tower kernel: ata4.00: READ LOG DMA EXT failed, trying PIO
  2. I am finally replacing 2 of my 4tb drives with 8tb drives and it really slows down at times (is this normal?) "Total size:8 TB Elapsed time:3 days, 15 hours, 51 minutes Current position:3.64 TB (45.4 %) Estimated speed:772.7 KB/sec Estimated finish:" If i have something like a power outage or a crash during this 6 to 8 day rebuild of both disks would i be able to use my old 4tb drives safely? or if it made it past the 4tb in the rebuild would my data be safe as it is since there are only zeros left to write (and they should already be zeros from preclear?) ( i have disabled mover so content on array should remain the same) Are parity calculations going to be done for the last 4tb writing zeros, or just write zeros? As far as i can tell there are is no change to parity happening during the data rebuild. At other times it peaks out at 20-23 MB/S and i anticipate the actual time to complete everything will be 6 to 8 days. I am leaving all my dockers running - Plex media server, sonarr radarr, sabnzbd, nzbhyrdra2. Should that have much effect on speed? This is all kind of scary, i have 3 more 4tb drives that i would like to replace with 8's though i would not mind reducing the array size a little in the process, and eventually the 5's to be replaced by 8's. Thanks in advance for answering any questions.
  3. From reading it sounds like just disable plugin restart and see if it goes away. I did, and deleted and reinstalled, error comes back.
  4. Yes, I have ups, we have really dirty power here though. Lots of storms, not too many unclean shutdowns since the UPS's were installed. I looked at my history on my other server and i had 1 to 4 errors fairly often.
  5. OK, i had 29 on that parity check and down to zero on this last one. I also ran another parity on my other server which had 3 errors on the last check and it came back zero. How much would ECC memory help? And if not that what would be your top recommendation to improve chances of less or no sync errors in future.
  6. I removed appdata from scanned directories and no longer have 100% cpu spikes. Wish i had known that a long time ago, hopefully my server will be more responsive without the spikes.
  7. 29 errors, doing one more, hopefully this is clean. thanks.
  8. I guess its good that its 12 errors at almost 10%, so instead of 5901 at 100% on last check. Is it expected to have to more than 1 more parity check to get back to zero if everything is ok?
  9. just shows under sync errors correct, all zeros in the errors column.
  10. I started memtest, was running fine for a while, but then decided to just swap out the box completely. - still using the same 16 bay expansion as before that is connected to the box through the same lsi 9207-8e and cable. No other parts are the same. The old box was an xps 8700 with 4770 intel cpu/8gb ram. - new one is same specs with 16gb of different ram. Running another parity check now, should be done in about 27 or so hours.
  11. I will run memtest when the parity check completes in about 10 hours. I can't say for certain that i have on this box, i did one in april prior to installing UPS's on both of my unraid servers. Before that i had a lot more problems with my servers. Installing UPS's helped a lot. We have quite a lot of outages here, 10 to 20 a year. At that point (and you had had helped me in that thread) i had about 300 errors on each servers after a few unclean shutdowns due to power outages/brownouts. This box in question has been running for about 60 days without a reboot, there have been a few brownouts/outages during that time frame, but long enough to tell unraid/ups to shut down the machine (less than 30 seconds). "you won't be able to reliably rebuild it if parity isn't completely correct." that means it will rebuild unreliable data? - I am currently at 3542 sync errors with about 33% left to go on parity check. I assume that means that i have some questionable data even once parity checks get back to no errors? If in the future i see any sync errors, just do another parity check? even if its 1 or 2?
  12. Is there a way to tell which drives have the sync errors (to determine if all on one drive)? I see log entries that have P corections, Q corrections and PQ corrections, but only sectors not any other indicator. example - Jan 1 00:05:12 PIPE kernel: md: recovery thread: PQ corrected, sector=42826928 Jan 1 00:05:19 PIPE kernel: md: recovery thread: PQ corrected, sector=43760832 Jan 1 00:05:27 PIPE kernel: md: recovery thread: PQ corrected, sector=44727040 Jan 1 00:05:31 PIPE kernel: md: recovery thread: Q corrected, sector=45261232 Jan 1 00:05:31 PIPE kernel: md: recovery thread: Q corrected, sector=45262752 Jan 1 00:05:51 PIPE kernel: md: recovery thread: P corrected, sector=47864688 Jan 1 00:05:51 PIPE kernel: md: recovery thread: PQ corrected, sector=47885344 Jan 1 00:05:59 PIPE kernel: md: recovery thread: PQ corrected, sector=48868
  13. LSI SAS HBA 9207-8e 6GBs 8-Port - i haven't had these in the past though, i added another 8tb drive recently. Thanks for the clarification on cache drive not being part of the problem.
  14. I have several of these, i can't see the whole screen so ill type it as best i can from a screenshot. Im not sure if t here is a letter before the xz and cat. Should i worry about it? I don't do any script writing. *xz: (sdin): Compressed data is corrupt cat: write error: Broken Pipe xz: (sdin): Compressed data is corrupt cat: write error: Broken Pipe cat: write error: Broken Pipe Currrently running 6.6.6 and all plugins are up to date - still shows up after each reboot.
  15. Hi Johnnie, I just saw this, bummed i missed it when you posted it. I always appreciate your advice. Happy New Year!
  16. Last month i had 660 sync errors, I'm on 1780 so far, 1/3 through this parity check. I have a udma crc error count 1 that intermittantly shows up on the cache drive - could this be causing it? I need to change the cache drive. I have clock unsyncronized errors in my log - assume that means i need to change cmos battery. What are the consequences of these sync errors, does that mean that the data was bad and it has been fixed? or that the data on those sectors is bad after the parity check? Can i Interrupt the monthly parity check, change the cache drive, troubleshoot etc, and restart the parity check? approx 20 hours to go on it. Update - i should specify, i have 2 servers, this is Pipe - the other problem posted was Tower, a different server. Can someone have a look at my logs and advise me how to proceed? Thanks in advance! pipe-diagnostics-20190101-1036.zip
  17. Thanks as always johnnie for your help. Are the smart attributes looking ok for the ssd?
  18. I noticed that my docker wasn't responding since last night. I have a 1tb crucial ssd for cache, maybe something wrong with cache? tried to download smart for ssd, but doesn't look like that worked. All dockers show that they need update. Diagnostics attached - going to power down now and reboot - log shows a bunch of problems starting last night. Will attach an updated smart for ssd if i can get it after a reboot . - edit update after reboot (it had to force a shutdown after starting via gui - attaching screenshot of tower and smart report for cache drive that wasn't included in diagnostics. Note - write errors in screenshot, these were present after i rebooted about a week ago after kernel panic that i posted last week with no reply. tower-diagnostics-20181223-1119.zip tower-smart-20181223-1144.zip
  19. Last night during parity check, approximately 3/4 way through a 1 day 6 hour check. It appears that the server (named tower) crashed(quite likely when the scheduled mover started) - appears to have killed my network or part of my network. I will attach a screenshot of local monitor. I was unable to type on the keyboard locally and due to network problems as well as crash(I assume) i couldn't telnet in. So all i have that i know of is the screenshot 1 (bottom). Also see screenshot 2(top one) of tower booted, with some strange can't write errors. I am also attaching diagnostics taken after first reboot. Another factor is that the my router was not assigning reserverd IP address to tower on 3 reboot tests. However when i powered down the router and rebooted, the following reboot of tower got the proper IP address. I apologize for the slightly blurry screenshot. tower-diagnostics-20181217-0801.zip
  20. My plan is to do this as per the unraid 6 manual - Replace a Single Disk with a Bigger One This is the case where you are replacing a single small disk with a bigger one: 1. Stop the array. 2. Power down the unit. 3. Replace smaller disk with new bigger disk. 4. Power up the unit. 5. Start the array I also plan on preclearing the new 8tb to be safe. This is a large array - i have 2 8tb parity 5 4tb 7 5tb 8 6tb 2 8tb data Ultimately, i would like to get rid of all the 4tb drives or at least 4 of them, just to have a amount of disks. Any thoughts? or advice. My main concern is a power outage or failure of a drive during rebuild. Even considering getting another license for unraid and starting a fresh 8tb only array, eventually pulling the 4 8tbs (2 parity and 2 data) out Would a rebuild of 4tb onto an 8 take just as long as rebuilding an 8tb parity drive does? This takes a long time, really long time. If i remember correctly it has to write zeros the whole time.
  21. From what i have read i should not set my ups to shut off, because it will or may not (Tripplite). Although it seems that setting it to shut off after the computer shuts down would be worth trying? Anyhow does this look good enough to just get my server shut off if the power goes out? I set it to 30 seconds and assume that will be what happens versus runtime or batt level? Is there a safe way to test this and see if it is working? maybe just start the server but not the array and unplug ups from wall? See attached screen shot of config. Thanks.
  22. Darn , i missed that window, my brother is just flying out now! - hope fully i can get some feedback from someone who has used this one specifically with that one - easy to get from costco and have a friend bring back. Scary stuff that the last link i shared though.
×
×
  • Create New...