Jax Posted October 3, 2019 Share Posted October 3, 2019 (edited) Hi, This all began after initiating a drive replacement due to errors I was receiving on disk 6. I installed a new disk (4TB to replace the 3TB that had failed) and it seemed to start OK, but when I woke up this morning to check, I see that it's going to take a year to complete at the current rate: I am also seeing that the log is pegged @100% only after 7 hours of system uptime: Syslog is loaded with a lot of REISERFS errors related to md6. (attached) Any ideas as to what could be going on here? Any help or direction provided would be greatly appreciated... thanks! Edited October 6, 2019 by Jax Issue solved Quote Link to comment
trurl Posted October 3, 2019 Share Posted October 3, 2019 How did you determine the original disk was bad? Go to Tools - Diagnostics and attach the complete diagnostics zip file to your next post. Quote Link to comment
Jax Posted October 3, 2019 Author Share Posted October 3, 2019 (edited) On 10/3/2019 at 2:25 PM, trurl said: How did you determine the original disk was bad? Couldn't be mounted or read in Unraid - unable to perform a SMART scan... now that I have it out I can do some more checking. I've had drives fail before and this didn't appear to be any different. So since I had the spare on hand, I swapped it out to ask questions later. Full diagnostic attached. Edited October 6, 2019 by Jax Quote Link to comment
trurl Posted October 3, 2019 Share Posted October 3, 2019 2 hours ago, Jax said: Couldn't be mounted or read in Unraid - unable to perform a SMART scan These could all be caused by a bad connection. But could be a bad disk of course. 2 hours ago, Jax said: So since I had the spare on hand, I swapped it out to ask questions later. Rebuilding to a spare is actually the best way even if the original is still good, since it keeps the original as a backup in case of problems. Looks like it's having problems communicating with multiple disks. Reseat controller, check connections, both ends, including power. Power splitters are also a good suspect with multiple disk problems. The filesystem problems might not be real if it can't read all the disks to calculate the rebuild. Maybe it will clear up if you get all the disks connected again. 1 Quote Link to comment
Jax Posted October 4, 2019 Author Share Posted October 4, 2019 Thanks. Should I just cancel the rebuild to power down and check the connections? It's still stuck at 48.2% for the past 10 hours, but I don't want to do something that could result in data loss. Quote Link to comment
trurl Posted October 4, 2019 Share Posted October 4, 2019 34 minutes ago, Jax said: Should I just cancel the rebuild to power down and check the connections? It's still stuck at 48.2% for the past 10 hours, but I don't want to do something that could result in data loss. No point in continuing with a rebuild that can't be producing the correct results, since all of the disks must be read reliably to reliably rebuild a disk. Quote Link to comment
JorgeB Posted October 5, 2019 Share Posted October 5, 2019 Looks like the typical SAS2LP problem, those controllers are not recommended for a long time, if possible replace them with LSI. Quote Link to comment
Jax Posted October 6, 2019 Author Share Posted October 6, 2019 Reseated all power and signal cables and all disks and the rebuild completed in a reasonable amount of time: Putting the original "failed" drive through it's paces on the bench - so far so good. Will also look into the SAS controller issue as I wasn't aware there was one..... thanks for the replies! Quote Link to comment
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.