optiman Posted September 8, 2021 Share Posted September 8, 2021 Anyone using the new WD 14tb RED Pro drives with LSI controller with success? I've never had 2 drives fail at the same time, very weird. Both are Iron Wolf 12tb and only 3 years old. SMART data and logs look good, so not sure why unraid disabled them. Diag file attached if anyone can help sort out what happened. Thanks! tower-diagnostics-20210907-0859.zip Quote Link to comment
trurl Posted September 8, 2021 Share Posted September 8, 2021 Looks like they were already disabled before you booted so can't see why they were disabled. Most of the time nothing wrong with the disks so don't know why your first instinct is to go shopping for different models. Connection/controller problems are far more common than bad disks. Corruption on both disk2 and cache. Leave cache out of it for now until you get your array stable again. Go to Settings and disable both Docker and VM Manager until everything is fixed. SMART for parity and disk2 looks OK. Emulated disk2 not mounted though due to corruption. Don't do anything without further advice. Do any of your other disks have SMART warnings on the Dashboard page? Quote Link to comment
optiman Posted September 8, 2021 Author Share Posted September 8, 2021 Thanks. No, all disks SMART logs are good and no errors. I will check the controller and all cables and power. How can I get unraid to recognize disk 2 and parity again to try again? Quote Link to comment
trurl Posted September 8, 2021 Share Posted September 8, 2021 First thing you want to do is try to repair the emulated disk2. https://wiki.unraid.net/Manual/Storage_Management#Running_the_Test_using_the_webGui Quote Link to comment
optiman Posted September 8, 2021 Author Share Posted September 8, 2021 I dl the diag file first thing in the morning before I rebooted, so I have the previous logs, see attached. tower-diagnostics-20210907-0859.zip Quote Link to comment
trurl Posted September 8, 2021 Share Posted September 8, 2021 Those are the same diagnostics you posted earlier Quote Link to comment
optiman Posted September 8, 2021 Author Share Posted September 8, 2021 dam, the file overwrote the first one, not my day. I will try to Repair disk 2 as you suggested Quote Link to comment
optiman Posted September 9, 2021 Author Share Posted September 9, 2021 Ok, ran the repair and this was the output Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... ALERT: The filesystem has valuable metadata changes in a log which is being destroyed because the -L option was used. - scan filesystem freespace and inode maps... sb_icount 34816, counted 480192 sb_ifree 243, counted 251 sb_fdblocks 245643713, counted 791827383 - found root inode chunk Phase 3 - for each AG... - scan and clear agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - agno = 4 - agno = 5 - agno = 6 - agno = 7 - agno = 8 - agno = 9 - agno = 10 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 0 - agno = 2 - agno = 1 - agno = 6 - agno = 5 - agno = 4 - agno = 3 - agno = 7 - agno = 8 - agno = 9 - agno = 10 Phase 5 - rebuild AG headers and trees... - reset superblock... Phase 6 - check inode connectivity... - resetting contents of realtime bitmap and summary inodes - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify and correct link counts... Maximum metadata LSN (2:880019) is ahead of log (1:2). Format log to cycle 5. done I did have to use the -L I then restarted and the drive was still disabled What should I try next? Quote Link to comment
trurl Posted September 9, 2021 Share Posted September 9, 2021 21 hours ago, trurl said: First thing you want to do is try to repair the emulated disk2. 1 hour ago, optiman said: the drive was still disabled Because you haven't done the next thing yet. The point of the repair was to get the emulated disk mountable before trying to rebuild it. Post new diagnostics. Quote Link to comment
optiman Posted September 9, 2021 Author Share Posted September 9, 2021 we lost power and my server auto shutdown when the ups got low. So I just booted it back up. I lost the system log. Next time I will dl the diag file right after each step. I'm guessing you don't want the current one, as I just booted up. What is the next step? Thank you so much for helping! Quote Link to comment
trurl Posted September 9, 2021 Share Posted September 9, 2021 53 minutes ago, optiman said: I'm guessing you don't want the current one yes I do Quote Link to comment
optiman Posted September 9, 2021 Author Share Posted September 9, 2021 diag file attached. I downloaded it right after I booted up the server. tower-diagnostics-20210909-1631.zip Quote Link to comment
trurl Posted September 10, 2021 Share Posted September 10, 2021 Emulated disk2 mounts thanks to the repair, it appears to be 74% full, and it doesn't look like anything actually got put in lost+found. You should be able to look at the files on the emulated disk. The contents of the emulated disk is exactly what will be rebuilt. Since you have dual parity you can rebuild both drives at the same time using the same disks. https://wiki.unraid.net/Manual/Storage_Management#Rebuilding_a_drive_onto_itself Quote Link to comment
optiman Posted September 10, 2021 Author Share Posted September 10, 2021 ok that's great to news! I've followed those instructions and it is rebuilding both disk2 and the parity drives right now. Once finished, I will post the diag file again and report back. Thank you!!! Quote Link to comment
optiman Posted September 10, 2021 Author Share Posted September 10, 2021 rebuild complete and it looks like everything is ok. Should I check or test anything else? I never did find any route cause as to why both of those drives were disabled. Diag file attached. The weird thing is it took forever to collect the diag info, so I hit Done and did it again, and it worked normal, took 10 seconds. I'll now try to sort out what is going on with my docker not starting. Thanks again! tower-diagnostics-20210910-1722.zip Quote Link to comment
optiman Posted September 10, 2021 Author Share Posted September 10, 2021 I tried to start the Docker, and again it says it cannot start. I went ahead and grabbed the diag file again so you can see that part in the log. tower-diagnostics-20210910-1725.zip Quote Link to comment
trurl Posted September 11, 2021 Share Posted September 11, 2021 Corrupt docker.img Delete and recreate https://wiki.unraid.net/Manual/Docker_Management#Re-Create_the_Docker_image_file Reinstall dockers https://wiki.unraid.net/Manual/Docker_Management#Re-Installing_Docker_Applications Quote Link to comment
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.