amixnux Posted February 11, 2023 Share Posted February 11, 2023 Hi, I had some problems with a Transmission Docker and I was unable to restart the Unraid, so I end-up with (re)start of my unraid. The system automatically started with a failing drive, but I did not realized. Transmission started to download files that supposed to be already in the mounted location, and then I realized that the the system had started completely wrong, missing one of Disks. I just started the array again without that disk in the list, so I though that way the content would be Virtualized from the Parity drive, but that did not work. I cancelled the parity check as yesterday I already did one. No sure how to proceed to recover my data. Attached the diagnostics thanks for any guidance ! unraid-diagnostics-20230211-1319.zip Quote Link to comment
itimpi Posted February 11, 2023 Share Posted February 11, 2023 Is disk1 showing as unmountable (which is what the diagnostics suggest)? If so the correct handling of unmountable disks is covered here in the online documentation accessible via the ‘Manual’ link at the bottom of the GUI or the DOCS link at the top of each forum page. Quote Link to comment
amixnux Posted February 11, 2023 Author Share Posted February 11, 2023 Thanks. I can mount it, I was not sure what I needed to do to repair it. Just added again to the array and it seems to work, even though I do not see the data in the shares Then all I need to do is wait? Quote Link to comment
trurl Posted February 11, 2023 Share Posted February 11, 2023 4 minutes ago, amixnux said: I can mount it You are apparently using that term incorrectly. Your screenshot says the disk is unmountable. You should repair the filesystem of the emulated disk. If that goes well then you will be able to see your files on the emulated disk. Then rebuild will write the emulated data to the drive. Start with 4 hours ago, itimpi said: handling of unmountable disks is covered here Be sure to capture the output so you can post it Quote Link to comment
trurl Posted February 11, 2023 Share Posted February 11, 2023 Wish you had waited for confirmation of what you were going to do, especially since you went in a different direction than the recommendation. I see from your screenshot you are already rebuilding the unmountable filesystem, so when it is finished you will still have an unmountable filesystem and won't be able to access any of its files. We usually prefer to repair the emulated filesystem before rebuilding, especially when rebuilding on top of the same disk. That way we have other options if things don't work out well. As it is, you will have to try to repair the filesystem on the rebuilt disk. Let us know when rebuild completes and we can work on that. Quote Link to comment
amixnux Posted February 11, 2023 Author Share Posted February 11, 2023 yes, I kind of new to this. Thanks for the clarification. Should I now stop the array, run the test, end then decide from there? Quote Link to comment
trurl Posted February 11, 2023 Share Posted February 11, 2023 Just now, amixnux said: Should I now stop the array, run the test, end then decide from there? Too late now. 4 minutes ago, trurl said: As it is, you will have to try to repair the filesystem on the rebuilt disk. Let us know when rebuild completes and we can work on that. Quote Link to comment
trurl Posted February 11, 2023 Share Posted February 11, 2023 Since you rebooted before getting those earlier diagnostics, no way to know what happened before to disable the disk. Did you do anything to possibly correct that, such as check the connections? Might be worth posting new diagnostics so we can see how the rebuild is going. Quote Link to comment
amixnux Posted February 11, 2023 Author Share Posted February 11, 2023 I was not counting with such a quick reply and help. Thank you Next time, I should take a break before doing anything in Panic and read more thoughtfully. I will post the results once done the rebuild Quote Link to comment
trurl Posted February 11, 2023 Share Posted February 11, 2023 2 minutes ago, trurl said: Might be worth posting new diagnostics so we can see how the rebuild is going. Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 current diagnostics, still two hours to finish unraid-diagnostics-20230212-0916.zip Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 It just finished successfully, . Now the disk shows green but still is not mountable. Diagnostics attached with the results unraid-diagnostics-20230212-1146.zip Quote Link to comment
itimpi Posted February 12, 2023 Share Posted February 12, 2023 Have you run the option mentioned earlier to check/repair the file system on the drive? Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 I did not, I want to make sure I get it right (and ask here) before making a process of 14hours that does not help Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 So, I should I stop the array then go to here disk and do this check? Following this " Repairing a File System" from https://wiki.unraid.net/Manual/Storage_Management#Drive_shows_as_unmountable ? Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 I stopped the array, started in Maintenance mode and started the test as the documentation suggests. I selected to not modify data. Phase 1 - find and verify superblock... - reporting progress in intervals of 15 minutes Phase 2 - using internal log - zero log... ALERT: The filesystem has valuable metadata changes in a log which is being ignored because the -n option was used. Expect spurious inconsistencies which may be resolved by first mounting the filesystem to replay the log. - 11:51:27: zeroing log - 30524 of 30524 blocks done - scan filesystem freespace and inode maps... agi unlinked bucket 23 is 12586199 in ag 6 (inode=817892567) sb_icount 406656, counted 406528 sb_ifree 32887, counted 32851 sb_fdblocks 896025097, counted 928275006 - 11:51:27: scanning filesystem freespace - 125 of 125 allocation groups done - found root inode chunk Phase 3 - for each AG... - scan (but don't clear) agi unlinked lists... - 11:51:27: scanning agi unlinked lists - 125 of 125 allocation groups done - process known inodes and perform inode discovery... - agno = 0 - agno = 15 - agno = 90 - agno = 30 - agno = 60 - agno = 45 - agno = 75 - agno = 105 - agno = 120 - agno = 91 - agno = 61 - agno = 31 - agno = 46 - agno = 76 - agno = 1 - agno = 32 - agno = 62 - agno = 77 - agno = 121 - agno = 92 - agno = 2 - agno = 47 - agno = 93 - agno = 106 - agno = 63 - agno = 33 - agno = 16 - agno = 78 - agno = 122 - agno = 48 - agno = 107 - agno = 123 - agno = 64 - agno = 3 - agno = 34 - agno = 124 - agno = 79 - agno = 49 - agno = 94 - agno = 17 - agno = 18 - agno = 19 - agno = 35 - agno = 65 - agno = 36 - agno = 80 - agno = 37 - agno = 38 - agno = 20 - agno = 4 - agno = 95 - agno = 66 - agno = 81 - agno = 67 - agno = 96 - agno = 97 - agno = 50 - agno = 68 - agno = 108 - agno = 51 - agno = 82 - agno = 39 - agno = 40 - agno = 21 - agno = 69 - agno = 41 - agno = 52 - agno = 98 - agno = 53 - agno = 83 - agno = 70 - agno = 42 - agno = 54 - agno = 109 - agno = 99 - agno = 43 - agno = 5 - agno = 22 - agno = 84 - agno = 55 - agno = 110 - agno = 100 - agno = 71 - agno = 44 - agno = 72 - agno = 56 - agno = 85 - agno = 57 - agno = 86 - agno = 101 - agno = 111 - agno = 58 - agno = 73 - agno = 23 - agno = 112 - agno = 113 - agno = 87 - agno = 74 - agno = 59 - agno = 88 - agno = 89 - agno = 6 - agno = 102 - agno = 24 - agno = 114 - agno = 103 - agno = 25 - agno = 115 bad fwd (right) sibling pointer (saw 102164487 parent block says 102642459) in inode 817892567 (data fork) bmap btree block 102236580 bad data fork in inode 817892567 would have cleared inode 817892567 - agno = 7 - agno = 26 - agno = 104 - agno = 27 - agno = 28 - agno = 8 - agno = 116 - agno = 117 - agno = 29 - agno = 9 - agno = 118 - agno = 10 - agno = 11 - agno = 119 - agno = 12 - agno = 13 - agno = 14 - 11:52:04: process known inodes and inode discovery - 406656 of 406656 inodes done - process newly discovered inodes... - 11:52:04: process newly discovered inodes - 125 of 125 allocation groups done Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - 11:52:04: setting up duplicate extent list - 125 of 125 allocation groups done - check for inodes claiming duplicate blocks... - agno = 0 - agno = 3 - agno = 2 - agno = 1 - agno = 4 - agno = 5 - agno = 6 - agno = 7 - agno = 8 - agno = 9 - agno = 10 - agno = 11 - agno = 12 - agno = 13 - agno = 14 - agno = 15 - agno = 16 - agno = 17 - agno = 18 - agno = 19 - agno = 20 - agno = 21 - agno = 22 - agno = 23 - agno = 24 - agno = 25 - agno = 26 - agno = 27 - agno = 28 - agno = 29 - agno = 30 - agno = 31 - agno = 32 - agno = 33 - agno = 34 - agno = 35 - agno = 36 - agno = 37 - agno = 38 - agno = 39 - agno = 40 - agno = 41 - agno = 42 - agno = 43 - agno = 44 - agno = 45 - agno = 46 - agno = 47 - agno = 48 - agno = 49 - agno = 50 - agno = 51 - agno = 52 - agno = 53 - agno = 54 - agno = 55 - agno = 56 - agno = 57 - agno = 58 - agno = 59 - agno = 60 - agno = 61 - agno = 62 - agno = 63 - agno = 64 - agno = 65 - agno = 66 - agno = 67 - agno = 68 - agno = 69 - agno = 70 - agno = 71 - agno = 72 - agno = 73 - agno = 74 - agno = 75 - agno = 76 - agno = 77 - agno = 78 - agno = 79 - agno = 80 - agno = 81 - agno = 82 - agno = 83 - agno = 84 - agno = 85 - agno = 86 - agno = 87 - agno = 88 - agno = 89 - agno = 90 - agno = 91 - agno = 92 - agno = 93 - agno = 94 - agno = 95 - agno = 96 - agno = 97 - agno = 98 - agno = 99 - agno = 100 - agno = 101 - agno = 102 - agno = 103 - agno = 104 Metadata CRC error detected at 0x451d6d, xfs_bmbt block 0x2da858f0/0x1000 btree block 6/1979162 is suspect, error -74 bad magic # 0x94b8bcbe in inode 817892567 (data fork) bmbt block 102642458 bad data fork in inode 817892567 would have cleared inode 817892567 - agno = 105 - agno = 106 - agno = 107 - agno = 108 - agno = 109 - agno = 110 - agno = 111 - agno = 112 - agno = 113 - agno = 114 - agno = 115 - agno = 116 - agno = 117 - agno = 118 - agno = 119 - agno = 120 - agno = 121 - agno = 122 - agno = 123 - agno = 124 - 11:52:04: check for inodes claiming duplicate blocks - 406656 of 406656 inodes done No modify flag set, skipping phase 5 Phase 6 - check inode connectivity... - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify link counts... - 11:52:37: verify and correct link counts - 125 of 125 allocation groups done Maximum metadata LSN (426063857:1465603830) is ahead of log (82:180372). Would format log to cycle 426063860. No modify flag set, skipping filesystem flush and exiting. unraid-diagnostics-20230212-1253.zip Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 should I run it again with the with no options so it does the repair? Quote Link to comment
Solution itimpi Posted February 12, 2023 Solution Share Posted February 12, 2023 1 minute ago, amixnux said: should I run it again with the with no options so it does the repair? Yes, and if it asks for it use the -L option Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 Thanks. I did it, it asked me for the -L option Results below. The drive still shows as unmountable Version: 6.11.5 Dashboard Main Shares Users Settings Plugins Apps Tools Disk 1 Settings Name: Disk 1 Identification: ST8000DM004-2CX188_ZR12SWSV (sdg) Comments: Partition size: 7,814,026,532 KB (K=1024) Partition format: GPT: 4KiB-aligned Spin down delay: File system status: Unmountable: Wrong or no file system File system type: Warning disk utilization threshold (%): Critical disk utilization threshold (%): Check Filesystem Status xfs_repair status: Phase 1 - find and verify superblock... - reporting progress in intervals of 15 minutes Phase 2 - using internal log - zero log... ALERT: The filesystem has valuable metadata changes in a log which is being destroyed because the -L option was used. - 13:22:06: zeroing log - 30524 of 30524 blocks done - scan filesystem freespace and inode maps... clearing needsrepair flag and regenerating metadata agi unlinked bucket 23 is 12586199 in ag 6 (inode=817892567) sb_icount 406656, counted 406528 sb_ifree 32887, counted 32851 sb_fdblocks 896025097, counted 928275006 - 13:22:06: scanning filesystem freespace - 125 of 125 allocation groups done - found root inode chunk Phase 3 - for each AG... - scan and clear agi unlinked lists... - 13:22:06: scanning agi unlinked lists - 125 of 125 allocation groups done - process known inodes and perform inode discovery... - agno = 15 - agno = 0 - agno = 105 - agno = 60 - agno = 120 - agno = 30 - agno = 90 - agno = 75 - agno = 45 - agno = 31 - agno = 46 - agno = 76 - agno = 47 - agno = 91 - agno = 92 - agno = 61 - agno = 121 - agno = 122 - agno = 106 - agno = 93 - agno = 77 - agno = 123 - agno = 62 - agno = 48 - agno = 32 - agno = 94 - agno = 49 - agno = 107 - agno = 1 - agno = 2 - agno = 50 - agno = 63 - agno = 78 - agno = 33 - agno = 16 - agno = 79 - agno = 95 - agno = 51 - agno = 108 - agno = 124 - agno = 64 - agno = 80 - agno = 3 - agno = 34 - agno = 52 - agno = 65 - agno = 17 - agno = 96 - agno = 97 - agno = 109 - agno = 35 - agno = 53 - agno = 81 - agno = 18 - agno = 4 - agno = 19 - agno = 66 - agno = 67 - agno = 36 - agno = 54 - agno = 82 - agno = 98 - agno = 83 - agno = 110 - agno = 68 - agno = 69 - agno = 55 - agno = 37 - agno = 84 - agno = 99 - agno = 111 - agno = 20 - agno = 21 - agno = 38 - agno = 5 - agno = 56 - agno = 70 - agno = 100 - agno = 112 - agno = 113 - agno = 85 - agno = 6 - agno = 22 - agno = 114 - agno = 115 - agno = 116 - agno = 57 - agno = 71 - agno = 101 - agno = 86 - agno = 87 - agno = 23 - agno = 39 - agno = 58 - agno = 102 - agno = 40 - agno = 72 - agno = 24 - agno = 117 - agno = 88 - agno = 41 - agno = 59 bad fwd (right) sibling pointer (saw 102164487 parent block says 102642459) in inode 817892567 (data fork) bmap btree block 102236580 bad data fork in inode 817892567 cleared inode 817892567 - agno = 7 - agno = 42 - agno = 25 - agno = 73 - agno = 89 - agno = 118 - agno = 103 - agno = 74 - agno = 43 - agno = 119 - agno = 104 - agno = 26 - agno = 44 - agno = 27 - agno = 8 - agno = 28 - agno = 29 - agno = 9 - agno = 10 - agno = 11 - agno = 12 - agno = 13 - agno = 14 - 13:22:42: process known inodes and inode discovery - 406656 of 406656 inodes done - process newly discovered inodes... - 13:22:42: process newly discovered inodes - 125 of 125 allocation groups done Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - 13:22:42: setting up duplicate extent list - 125 of 125 allocation groups done - check for inodes claiming duplicate blocks... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - agno = 4 - agno = 5 - agno = 6 - agno = 7 - agno = 8 - agno = 9 - agno = 10 - agno = 11 - agno = 12 - agno = 13 - agno = 14 - agno = 15 - agno = 16 - agno = 17 - agno = 18 - agno = 19 - agno = 20 - agno = 21 - agno = 22 - agno = 23 - agno = 24 - agno = 25 - agno = 26 - agno = 27 - agno = 28 - agno = 29 - agno = 30 - agno = 31 - agno = 32 - agno = 33 - agno = 34 - agno = 35 - agno = 36 - agno = 37 - agno = 38 - agno = 39 - agno = 40 - agno = 41 - agno = 42 - agno = 43 - agno = 44 - agno = 45 - agno = 46 - agno = 47 - agno = 48 - agno = 49 - agno = 50 - agno = 51 - agno = 52 - agno = 53 - agno = 54 - agno = 55 - agno = 56 - agno = 57 - agno = 58 - agno = 59 - agno = 60 - agno = 61 - agno = 62 - agno = 63 - agno = 64 - agno = 65 - agno = 66 - agno = 67 - agno = 68 - agno = 69 - agno = 70 - agno = 71 - agno = 72 - agno = 73 - agno = 74 - agno = 75 - agno = 76 - agno = 77 - agno = 78 - agno = 79 - agno = 80 - agno = 81 - agno = 82 - agno = 83 - agno = 84 - agno = 85 - agno = 86 - agno = 87 - agno = 88 - agno = 89 - agno = 90 - agno = 91 - agno = 92 - agno = 93 - agno = 94 - agno = 95 - agno = 96 - agno = 97 - agno = 98 - agno = 99 - agno = 100 - agno = 101 - agno = 102 - agno = 103 - agno = 104 - agno = 105 - agno = 106 - agno = 107 - agno = 108 - agno = 109 - agno = 110 - agno = 111 - agno = 112 - agno = 113 - agno = 114 - agno = 115 - agno = 116 - agno = 117 - agno = 118 - agno = 119 - agno = 120 - agno = 121 - agno = 122 - agno = 123 - agno = 124 - 13:22:43: check for inodes claiming duplicate blocks - 406656 of 406656 inodes done Phase 5 - rebuild AG headers and trees... - 13:22:49: rebuild AG headers and trees - 125 of 125 allocation groups done - reset superblock... Phase 6 - check inode connectivity... - resetting contents of realtime bitmap and summary inodes - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify and correct link counts... - 13:23:24: verify and correct link counts - 125 of 125 allocation groups done Maximum metadata LSN (82:180369) is ahead of log (1:2). Format log to cycle 85. done Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 Should I now stop the array and start it once more in normal mode, as the next step from the documentation suggest. It finished successfully as far as I understand it, but worth to ask. Thank you itimpi for your help! Quote Link to comment
amixnux Posted February 12, 2023 Author Share Posted February 12, 2023 I started the array normally and like magic, all the things seems to be back. Thanks you so much @itimpi !! For any other person having a similar issue. 1) Calm down 2) Read the documentation from here; https://wiki.unraid.net/Manual/Storage_Management#Drive_shows_as_unmountable Start with the tests for the disks, 3) Ask the forum if something is not clear. Quote Link to comment
itimpi Posted February 12, 2023 Share Posted February 12, 2023 2 hours ago, amixnux said: Should I now stop the array and start it once more in normal mode, as the next step from the documentation suggest. It finished successfully as far as I understand it, but worth to ask. Thank you itimpi for your help! The one thing as a final check is to see if you have a Lost+Found folder (which will then appear as a share of that name). That is where the repair process puts any files for which it cannot find the directory information. Not having such a folder is always a GOOD sign as that suggests nothing fell into that category and sorting out the folders/files there without their names can be laborious (and often not worth the effort). Quote Link to comment
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.