gerhard911

Members
  • Posts

    53
  • Joined

  • Last visited

Everything posted by gerhard911

  1. I have replaced / rebuilt a data disk three times now and after each parity sync which appears to run successfully I end up with a message after starting the array that the disk is unmountable with no file system. This all started with data disk5 showing a lot of read errors. It was an older 8TB HGST drive so I swapped in a newer Segate 8TB and synched parity. That rebuild worked fine but the new drive also experienced read errors. Having planned to increase this server's capacity anyway, I purchased a pair of Segate 14TB drives and replaced the 10TB parity drive with a 14TB drive and rebuilt parity. That appeared to run without issues so I swapped out the Seagate 8TB disk5 for the previous parity 10TB drive and rebuilt data. That resulted in my first unmountable disk. I stopped the array, unassigned the 10TB disk and moved it to a previously unused SATA port. Restart the server, assign the 10TB disk to disk5 and rebuild data. Again, after appearing to rebuild successfully I end up with an unmountable drive. So I swap the SATA port & cable between disk5 and a cache pool drive and rebuild a third time. Same result with disk5 (unmountable) but now the cache drive is also unmountable. I downloaded diagnostics after each step and am attaching those .zip files. I also took screen captures of the Main page after each step in the parity sync / restart process and will attach those .jpg files. I realize now I should have probably saved the original 10TB parity drive intact and rebuilt disk5 on the other new 14TB. I am concerned that the unmountable disk issue is due to something happening with the 14TB parity drive swap. So I am tempted to rebuild parity with the original 8TB drive but need to know if I can trust the original 8TB disk5's data. Is there a way I can check that drive to see if it's data is intact? I have the Unassigned Devices plugin but don't know if / how it can be used to do this. Many thanks for any advice or assistance. unraid2-diagnostics-Parity sync done.zip unraid2-diagnostics-Parity sync stopped.zip unraid2-diagnostics-Post parity sync unmountable.zip
  2. OK, thanks for the followup. We are assuming that even though disk13 had a bunch of read errors and got disabled, that the data on it will be fine if it mounts with UD in read only mode.
  3. OK, I apologize I don't know what 'with UD in read-only mode' means. Is that starting the array in 'Maintenance Mode'? How do I confirm it has mounted correctly? I'm also a bit confused about the details of the new config process. Please bear with me and confirm I have this right. I replace the improperly rebuilt 14TB with the original 10TB. Then I do a New Config using what 'Preserve current assignments' ? It seems I would check both Parity slots and Data slots. I follow the New Config by rebuilding data on disk6. Is this what you referred to as a parity resync? Following that I can reattempt my 10TB to 14TB upgrade. Thanks for your expertise!
  4. After reboot disk13 still disabled. I did not restart the array. I know my original disk6 has all data intact and parity should be in sync with that disk. Again, is there a way (new config ?) to rebuild disk13 from parity with my original disk6 ? unraid1-diagnostics-reboot-20200907-1104.zip
  5. Here is the syslog during the rebuild. unraid1-14TB rebuild disk errors-syslog-20200904-1817.zip
  6. unraid1-diagnostics-disk13 disabled-20200907-1014.zip
  7. I was upgrading a 10TB disk to a 14TB disk (#6) and during the rebuild hundreds of thousands of read errors were reported on a 4TB disk (#13). The rebuild seemed to complete successfully but I can see that the rebuilt drive is missing most of the data it should have. Additionally, disk13 is now in a disabled state. I have the original 10TB disk6 intact. No data has been written to the array since the rebuild so parity should reflect the state of the array before rebuilding. So it seems to me that there should be some way to put the original disk6 back in and tell unRaid to rebuild disk13. Can this be accomplished or am I going to lose data? TIA for any advice or assistance.
  8. I have now been running on dual power supplies for several days. I have started and stopped the array several times without any unclean shutdowns. Thanks to Constructor I'm going to mark this thread 'solved'.
  9. So after the parity check completed, I carefully spun up the drives one at a time before stopping the array. All good! I attempted to replace the PS with an Antec 620W but because HP, the PCIe MB connector was non standard. So I hot wired a 480W Seasonic and am running dual PS. The MB, 1x10TB and 2 cache drives on the HP PS and all of the other drives on the Seasonic (5 each on two different rails). Fingers crossed.
  10. The server is an HP Proliant ML150G6. The power supply is a Delta Electronics DPS-480DB-2A. Specs listed on the label are Max power 480W at 12Va and Max Combined power 385W. 10 data drives (6x8 + 4x10TB) 2x1TB SSD cache and 10TB parity. Guess I need to check power requirements and add them all up. Thanks
  11. I have an unRAID server that has been plagued with unclean shutdowns for quite a while now. Initially it would only happen occasionally when a seemingly normal shutdown was performed. Now it happens with every shutdown. I have scoured syslogs finding nothing abnormal to my eye. I have also tried to determine a pattern of circumstance. This led me to suspect it was due to open files but using the Open Files tool to close suspect files would often cause an unclean shutdown itself. Making sure there were no open files from the client end did not help either. So today I had just completed a replace / rebuild of a data drive and clicked on the 'done' button. I then clicked on 'spin up all drives' and boom it triggered an unclean shutdown. syslog attached plus a screen capture of the flash drive showing no diagnostics log file. TIA for any advice or assistance. unraid2-mystery stop-syslog-20200901-1400.zip
  12. Looking at my server's Main page, I just realized that the "All existing data on this device will be OVERWRITTEN when array is Started" warning appears to be associated with just my parity drive. I had been spooked that it meant that all data on the whole array would be overwritten. If I start my server as is, with 20 data drives can someone confirm that parity will be rebuilt based on the data on those drives? TIA
  13. OK, I may have messed up in my attempt to shrink my 23 drive array down to 20. I cleared all data from the 3 drives I wanted to remove by moving it to two drives I was retaining that had sufficient space. I then physically removed the three drives and went to Tools and did a New Config with the default of 'retain none'. I had all of my drive assignments so I manually defined those on the Main page moving data drives up in the order to replace those removed. I now have parity and 20 data drives assigned but there is a warning that "All existing data on this device will be OVERWRITTEN when array is Started" ! I now realize that I missed the step of unassigning the three drives I was removing after doing a New Config and that I should have selected the 'retain all' option. So I'm pretty sure I don't want to start my array as it stands based on the warning but I do have the three drives that I removed and all of my data resides on the other 20 drives. Is there any way to get my array restarted, retain my data and rebuild parity? Ver 6.6.6 by the way. Thanks in advance for any help.
  14. Thanks for confirming my understanding of the New Configuration tool. I have parity rebuilding now with the original disk 21. Unfortunately that will take ~ a day and a half. I also appreciate the SATA / USB docking station suggestion. My experience with those units, however has been poor. I have owned and used several, from multiple brands. Every single one eventually failed, some quite quickly. I suspected some cheap common chipset or other component was used in them all and gave up on the whole concept. I should put in a single port external SATA controller card so I can attach an external drive bay for such purpose in the future. Thanks again for your help itimpi !
  15. I do still have the original 3TB drive but no free SATA port to attach it to. If I swap it back in as disk 21, I believe using the New Configuration tool will rebuild parity from data on all 23 disks. Is that correct?
  16. So, I had just swapped in a 4TB replacement for an aging 3TB drive and rebuilt my 23 drive array. I always rebuild in maintenance mode, so after the rebuild appeared to complete successfully, I shut down unraid (v 6.3.5) and started the array back up. Unraid flagged disk 21 (the one that had just been swapped and rebuilt) as "unmountable", saying it needed to be formatted. I let unraid format the disk (hopefully not a mistake) and now the array is started with valid configuration but no data on disk 21. Why doesn't unraid know that data is missing from that disk and want to rebuild from parity? Did the process of formatting disk 21 cause parity to drop all information for that disk or can I possibly remove that disk from the array, add it back and attempt a data rebuild from parity? TIA for any help or advice.
  17. Thanks itimpi. You have given me hope all is not lost with this drive. I am cloning the drive and wading through search results now.
  18. Doh! I did not pick up on the difference and did run the super block rebuild on /dev/sdh. Am I screwed on recovering any data from this drive?
  19. I have a serious problem with my array and am asking for advice or assistance from the experts here. And no, stupidly, I did not capture any syslogs. I had a relatively new (just a few weeks) drive red ball while saving a video file. Being fairly confident that the drive was OK, I stopped the array, unassigned/reassigned the drive, restarted the array and the instant rebuild started a second drive redballed! I ran reiserfsck --check on the second drive and it told me that the super block was corrupted and I should run reiserfsck --rebuild-sb. I did some research and found this thread: http://lime-technology.com/forum/index.php?topic=1483 I followed the instructions and the process appeared to complete although it issued no messages. I then reran reiserfsck --check with the following result: reiserfsck --check started at Mon Mar 24 17:39:56 2014 ########### Replaying journal: Done. Reiserfs journal '/dev/sdh' in blocks [18..8211]: 0 transactions replayed Zero bit found in on-disk bitmap after the last valid bit. Checking internal tree.. Bad root block 0. (--rebuild-tree did not complete) I am curious about the last line concerning --rebuild-tree since I did not run it with that option. Should I? The second failed drive is a 3TB which was almost completely full. I would really like to recover as much from it as possible. The first drive probably only had a few hundred GB on it and I am resigned that there is probably no way to recover data from it since a parity rebuild had started. Is my worst case scenario doing a "new configuration" without either of these drives? Thanks in advance for any help.
  20. Thanks for the response, Joe. I restarted the second preclear (it was still in pre-read) using v1.13. So where my first preclear ended with "1 sector is pending re-allocation at the end of the preclear" I am looking for the next run to show "0 sectors pending re-allocation at the end of the preclear". Fingers crossed. edit: I just realized I indicated I upgraded to unRaid 5.0-RC4. I have no idea where that came from, it was RC8a. I did not try any of the previous release candidates.
  21. After upgrading my server from beta 13 to RC4 (edit-it was 5.0-RC8a) I ran a parity check. No sync errors were found but immediately after the check finished unRaid red balled disk 7 (of nine). I left the disk in the server but removed it from the array, rebooted the server and restarted the array with disk 7 unassigned. The array came up in unprotected mode and looked fine from a data standpoint. I stopped the array and attempted to reassign the original disk 7 but it showed as unformatted and unRaid had the Format option greyed out. So I decided to run a preclear on the disk (it had been precleared before being initially added to the array). Here are the results: ============================================================= 1.12 = unRAID server Pre-Clear disk /dev/sde = cycle 1 of 1, partition start on sector 1 = Disk Pre-Clear-Read completed DONE = Step 1 of 10 - Copying zeros to first 2048k bytes DONE = Step 2 of 10 - Copying zeros to remainder of disk to clear it DONE = Step 3 of 10 - Disk is now cleared from MBR onward. DONE = Step 4 of 10 - Clearing MBR bytes for partition 2,3 & 4 DONE = Step 5 of 10 - Clearing MBR code area DONE = Step 6 of 10 - Setting MBR signature bytes DONE = Step 7 of 10 - Setting partition 1 to precleared state DONE = Step 8 of 10 - Notifying kernel we changed the partitioning DONE = Step 9 of 10 - Creating the /dev/disk/by* entries DONE = Step 10 of 10 - Verifying if the MBR is cleared. DONE = Disk Post-Clear-Read completed DONE Disk Temperature: 37C, Elapsed Time: 62:40:10 ========================================================================1.12 == WDC WD30EZRX-00MMMB0 WD-WCAWZ2393453 == Disk /dev/sde has been successfully precleared == with a starting sector of 1 ============================================================================ ** Changed attributes in files: /tmp/smart_start_sde /tmp/smart_finish_sde ATTRIBUTE NEW_VAL OLD_VAL FAILURE_THRESHOLD STATUS RAW_VALUE Raw_Read_Error_Rate = 172 200 51 ok 28610 Spin_Up_Time = 253 225 21 ok 3108 Seek_Error_Rate = 100 200 0 ok 0 Temperature_Celsius = 115 116 0 ok 37 No SMART attributes are FAILING_NOW 0 sectors were pending re-allocation before the start of the preclear. 6 sectors were pending re-allocation after pre-read in cycle 1 of 1. 0 sectors were pending re-allocation after zero of disk in cycle 1 of 1. 1 sector is pending re-allocation at the end of the preclear, a change of 1 in the number of sectors pending re-allocation. 0 sectors had been re-allocated before the start of the preclear. 0 sectors are re-allocated at the end of the preclear, the number of sectors re-allocated did not change. I am concerned about the sectors pending reallocation. I have a second preclear running now but wanted to get some feedback from those with more knowledge. TIA
  22. Not sure what model ReadyNAS you have but neither of mine have an eSATA port. If yours does it's probably intended to support an additional drive, not connect to a host system.
  23. Thank you! Hopefully the telnet console will do the trick.
  24. My unraid server has crashed several times with this error. From the last console screen I can see the error message in the title repeated multiple times. Process can be hdparm or sh. Does this message indicate a Linux bug? Possibly a hardware issue? After each failure the server reboots and unraid starts the array without problems but a parity check auto starts (and runs for @ 16 hrs with a 3TB parity drive + 12 data drives). Server will generally run for a couple of days without issue. No discernible pattern to the crashes. Latest one I was writing to the array but it has happened overnight with no activity as well. There are also some other diagnostics that would be very time consuming to write down and possibly of little use without a full syslog. It appears that there is no syslog retained on the flash drive. Is there a way to force the syslog to write to flash? If so, do I need to be concerned about filling up the flash drive (2GB)? Thanks in advance for any assistance. edit to add I am running unRAID 5.0-beta13