Jump to content

pickthenimp

Members
  • Content Count

    169
  • Joined

  • Last visited

Community Reputation

1 Neutral

About pickthenimp

  • Rank
    Advanced Member

Converted

  • Gender
    Undisclosed

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. Thanks for the response. One time thing...so far. Was really hoping to do a clean powerdown so I wouldn't have to do a full parity check. Oh well. Does my theory sound plausible? All of my plugins were autoupdating at the time of a preclear and I was using the unmanaged devices plugin to pre-clear and I wonder if those updating caused my box to shit itself.
  2. WebUI became unresponsive today. I would try to get in via putty but after putting in my user name nothing would happen. I was able to get in via telnet and grab a syslog (attached). I attempted to reboot via 'reboot' and 'powerdown -r' both of which said "System is going down for a reboot now" but never actually rebooted. Now when I telnet it is just a black screen, no prompt for credentials. Any idea what could be going on before I hard power cycle? I was using it fine the day before. Seems like something didnt cleanly start back up after my backup job ran... Also, I was running a preclear at the time. Since the backup script seems to kill all dockers and plugins, maybe this was part of my problem? syslog.zip
  3. Thanks. I’m sure the log will spew again and I will check it out in a few days. Sent from my iPhone using Tapatalk
  4. du -ah /var/lib/docker/containers/ | grep -v "/$" | sort -rh | head -60 | grep .log I found the above command which tells me I have a 8.5 GB log somewhere in my docker: 8.5G /var/lib/docker/containers/651070b829a5e773eb905534612c7c3216f7b7003b93c50a35869b458a556002/651070b829a5e773eb905534612c7c3216f7b7003b93c50a35869b458a556002-json.log I did a 'rm' to remove this log file. Is this normal and will have to keep an eye out in the future?
  5. I got an alert my docker image utilization is at 100%. All of my dockers are configured to map folders to the cache disk. I fired up CADvisor and all of my dockers disk usage doesn't add up anywhere near to the 20GB I have provisioned. What else could be using all of the space?
  6. Good man. its back up and seems to be in good shape. Now just hope it doesnt happen again?
  7. thanks for the reply @johnnie.black I ran this the next time and got a different notice. Should I run xfs_repair -L /dev/sdb1 now? root@Tower:/home# xfs_repair -v /dev/sdb1 Phase 1 - find and verify superblock... - block cache size set to 758600 entries Phase 2 - using internal log - zero log... zero_log: head block 74016 tail block 70932 ERROR: The filesystem has valuable metadata changes in a log which needs to be replayed. Mount the filesystem to replay the log, and unmount it before re-running xfs_repair. If you are unable to mount the filesystem, then use the -L option to destroy the log and attempt a repair. Note that destroying the log may cause corruption -- please attempt a mount of the filesystem before doing this.
  8. Posting for a friend. Shortly after upgrading to 6.3.5 his dockers went offline and the cache drive is listed as unmountable. We attempted a disk repair and this is the result: xfs_repair status: Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... - scan filesystem freespace and inode maps... Metadata corruption detected at xfs_agf block 0x1/0x200 flfirst 118 in agf 0 too large (max = 118) agf 118 freelist blocks bad, skipping freelist scan agi unlinked bucket 23 is 7447 in ag 0 (inode=7447) agi unlinked bucket 57 is 8249 in ag 0 (inode=8249) agi unlinked bucket 59 is 42024699 in ag 0 (inode=42024699) sb_icount 163776, counted 164800 sb_ifree 1803, counted 1547 sb_fdblocks 16453555, counted 16027419 - found root inode chunk Phase 3 - for each AG... - scan (but don't clear) agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 2 - agno = 0 - agno = 3 - agno = 1 No modify flag set, skipping phase 5 Phase 6 - check inode connectivity... - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... disconnected inode 7447, would move to lost+found disconnected inode 8249, would move to lost+found disconnected inode 42024699, would move to lost+found Phase 7 - verify link counts... would have reset inode 7447 nlinks from 0 to 1 would have reset inode 8249 nlinks from 0 to 1 would have reset inode 42024699 nlinks from 0 to 1 No modify flag set, skipping filesystem flush and exiting. Attached are his diagnostics. Assuming we need to reformat the disk and try again, how would I recover the data if the drive is unmountable? tower-diagnostics-20170918-1728.zip
  9. I found this drive in my old drobo. The precelar results do not give me warm and fuzzies. What do you all think? ############################################################################################################################ # # # unRAID Server Preclear of disk WD-WMAZA2814809 # # Cycle 1 of 1, partition start on sector 64. # # # # # # Step 1 of 5 - Pre-read verification: [0:58:38 @ 92 MB/s] SUCCESS # # Step 2 of 5 - Zeroing the disk: [6:52:49 @ 80 MB/s] SUCCESS # # Step 3 of 5 - Writing unRAID's Preclear signature: SUCCESS # # Step 4 of 5 - Verifying unRAID's Preclear signature: SUCCESS # # Step 5 of 5 - Post-Read verification: [1:56:19 @ 87 MB/s] SUCCESS # # # # # # # # # # # # # # # ############################################################################################################################ # Cycle elapsed time: 9:47:51 | Total elapsed time: 9:47:51 # ############################################################################################################################ ############################################################################################################################ # # # S.M.A.R.T. Status default # # # # # # ATTRIBUTE INITIAL CYCLE 1 STATUS # # 5-Reallocated_Sector_Ct 0 335 Up 335 # # 9-Power_On_Hours 39552 39562 Up 10 # # 194-Temperature_Celsius 18 17 Down 1 # # 196-Reallocated_Event_Count 0 126 Up 126 # # 197-Current_Pending_Sector 1428 1 Down 1427 # # 198-Offline_Uncorrectable 1261 1261 - # # 199-UDMA_CRC_Error_Count 0 0 - # # # # # # # # # # # ############################################################################################################################ # SMART overall-health self-assessment test result: PASSED # ############################################################################################################################ --> ATTENTION: Please take a look into the SMART report above for drive health issues. --> RESULT: Preclear Finished Successfully!. root@Tower:/usr/local/emhttp# tower-smart-20170321-0956.zip
  10. I ended up getting that Marvell sata card anyway since I just need a quick and dirty fix while I wait on a new build I am purchasing. I rebuilt drive 4 successfully per @trurl recommendation. Running a parity sync now (without correction) with zero errors... Thanks everyone for the help.
  11. Thanks for the reply. Do you have a better controller you recommend?
  12. Thanks for digging in. As a quick solution, would purchasing a 4 port Sata controller and moving these 3 disks off the onboard controller solve my problem? Looking at buying this: https://www.amazon.com/gp/product/B00AZ9T3OU
  13. Via unc path directly to the disk from another machine. Sent from my iPhone using Tapatalk
  14. Yes these are all on the same onboard controller. But seems odd I can access all of those disks fine? Sent from my iPhone using Tapatalk