pickthenimp

Members
  • Posts

    169
  • Joined

  • Last visited

Everything posted by pickthenimp

  1. Thanks for the response. One time thing...so far. Was really hoping to do a clean powerdown so I wouldn't have to do a full parity check. Oh well. Does my theory sound plausible? All of my plugins were autoupdating at the time of a preclear and I was using the unmanaged devices plugin to pre-clear and I wonder if those updating caused my box to shit itself.
  2. WebUI became unresponsive today. I would try to get in via putty but after putting in my user name nothing would happen. I was able to get in via telnet and grab a syslog (attached). I attempted to reboot via 'reboot' and 'powerdown -r' both of which said "System is going down for a reboot now" but never actually rebooted. Now when I telnet it is just a black screen, no prompt for credentials. Any idea what could be going on before I hard power cycle? I was using it fine the day before. Seems like something didnt cleanly start back up after my backup job ran... Also, I was running a preclear at the time. Since the backup script seems to kill all dockers and plugins, maybe this was part of my problem? syslog.zip
  3. Thanks. I’m sure the log will spew again and I will check it out in a few days. Sent from my iPhone using Tapatalk
  4. du -ah /var/lib/docker/containers/ | grep -v "/$" | sort -rh | head -60 | grep .log I found the above command which tells me I have a 8.5 GB log somewhere in my docker: 8.5G /var/lib/docker/containers/651070b829a5e773eb905534612c7c3216f7b7003b93c50a35869b458a556002/651070b829a5e773eb905534612c7c3216f7b7003b93c50a35869b458a556002-json.log I did a 'rm' to remove this log file. Is this normal and will have to keep an eye out in the future?
  5. I got an alert my docker image utilization is at 100%. All of my dockers are configured to map folders to the cache disk. I fired up CADvisor and all of my dockers disk usage doesn't add up anywhere near to the 20GB I have provisioned. What else could be using all of the space?
  6. Good man. its back up and seems to be in good shape. Now just hope it doesnt happen again?
  7. thanks for the reply @johnnie.black I ran this the next time and got a different notice. Should I run xfs_repair -L /dev/sdb1 now? root@Tower:/home# xfs_repair -v /dev/sdb1 Phase 1 - find and verify superblock... - block cache size set to 758600 entries Phase 2 - using internal log - zero log... zero_log: head block 74016 tail block 70932 ERROR: The filesystem has valuable metadata changes in a log which needs to be replayed. Mount the filesystem to replay the log, and unmount it before re-running xfs_repair. If you are unable to mount the filesystem, then use the -L option to destroy the log and attempt a repair. Note that destroying the log may cause corruption -- please attempt a mount of the filesystem before doing this.
  8. Posting for a friend. Shortly after upgrading to 6.3.5 his dockers went offline and the cache drive is listed as unmountable. We attempted a disk repair and this is the result: xfs_repair status: Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... - scan filesystem freespace and inode maps... Metadata corruption detected at xfs_agf block 0x1/0x200 flfirst 118 in agf 0 too large (max = 118) agf 118 freelist blocks bad, skipping freelist scan agi unlinked bucket 23 is 7447 in ag 0 (inode=7447) agi unlinked bucket 57 is 8249 in ag 0 (inode=8249) agi unlinked bucket 59 is 42024699 in ag 0 (inode=42024699) sb_icount 163776, counted 164800 sb_ifree 1803, counted 1547 sb_fdblocks 16453555, counted 16027419 - found root inode chunk Phase 3 - for each AG... - scan (but don't clear) agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 2 - agno = 0 - agno = 3 - agno = 1 No modify flag set, skipping phase 5 Phase 6 - check inode connectivity... - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... disconnected inode 7447, would move to lost+found disconnected inode 8249, would move to lost+found disconnected inode 42024699, would move to lost+found Phase 7 - verify link counts... would have reset inode 7447 nlinks from 0 to 1 would have reset inode 8249 nlinks from 0 to 1 would have reset inode 42024699 nlinks from 0 to 1 No modify flag set, skipping filesystem flush and exiting. Attached are his diagnostics. Assuming we need to reformat the disk and try again, how would I recover the data if the drive is unmountable? tower-diagnostics-20170918-1728.zip
  9. I found this drive in my old drobo. The precelar results do not give me warm and fuzzies. What do you all think? ############################################################################################################################ # # # unRAID Server Preclear of disk WD-WMAZA2814809 # # Cycle 1 of 1, partition start on sector 64. # # # # # # Step 1 of 5 - Pre-read verification: [0:58:38 @ 92 MB/s] SUCCESS # # Step 2 of 5 - Zeroing the disk: [6:52:49 @ 80 MB/s] SUCCESS # # Step 3 of 5 - Writing unRAID's Preclear signature: SUCCESS # # Step 4 of 5 - Verifying unRAID's Preclear signature: SUCCESS # # Step 5 of 5 - Post-Read verification: [1:56:19 @ 87 MB/s] SUCCESS # # # # # # # # # # # # # # # ############################################################################################################################ # Cycle elapsed time: 9:47:51 | Total elapsed time: 9:47:51 # ############################################################################################################################ ############################################################################################################################ # # # S.M.A.R.T. Status default # # # # # # ATTRIBUTE INITIAL CYCLE 1 STATUS # # 5-Reallocated_Sector_Ct 0 335 Up 335 # # 9-Power_On_Hours 39552 39562 Up 10 # # 194-Temperature_Celsius 18 17 Down 1 # # 196-Reallocated_Event_Count 0 126 Up 126 # # 197-Current_Pending_Sector 1428 1 Down 1427 # # 198-Offline_Uncorrectable 1261 1261 - # # 199-UDMA_CRC_Error_Count 0 0 - # # # # # # # # # # # ############################################################################################################################ # SMART overall-health self-assessment test result: PASSED # ############################################################################################################################ --> ATTENTION: Please take a look into the SMART report above for drive health issues. --> RESULT: Preclear Finished Successfully!. root@Tower:/usr/local/emhttp# tower-smart-20170321-0956.zip
  10. I ended up getting that Marvell sata card anyway since I just need a quick and dirty fix while I wait on a new build I am purchasing. I rebuilt drive 4 successfully per @trurl recommendation. Running a parity sync now (without correction) with zero errors... Thanks everyone for the help.
  11. Thanks for the reply. Do you have a better controller you recommend?
  12. Thanks for digging in. As a quick solution, would purchasing a 4 port Sata controller and moving these 3 disks off the onboard controller solve my problem? Looking at buying this: https://www.amazon.com/gp/product/B00AZ9T3OU
  13. Via unc path directly to the disk from another machine. Sent from my iPhone using Tapatalk
  14. Yes these are all on the same onboard controller. But seems odd I can access all of those disks fine? Sent from my iPhone using Tapatalk
  15. I double checked all my sata cables and they are snug Here is the latest: Upgraded to 6.3.2. Rebuilt disk 4 with a new drive. Reboot and missing shares finally came back Started a parity sync (without writing corrections) and it started out painfully slow. My error count kept going up. I finally stopped it after find 666670 errors. Latest diagnostics attached after stopping parity sync. Bad mobo? nas-diagnostics-20170302-0652.zip
  16. So, I got anxious and went out and purchased a new 2TB drive to replace disk4. Things locked up after I started the rebuild but I copied off the syslog. See attached. Any guidance would be appreciated. **Edit Unraid became responsive again and the drive seems to be rebuilding, however I am all of the sudden missing my TV and Movies shares?? Not sure what is going on.. About 2 minutes into the rebuild I got a few alerts: Event: unRAID array errorsSubject: Warning [NAS] - array has errorsDescription: Array has 1 disk with read errorsImportance: warningDisk 3 - WDC_WD20EARS-00MVWB0_WD-WMAZA3690347 (sdj) (errors 48) Event: unRAID array errors Subject: Warning [NAS] - array has errors Description: Array has 2 disks with read errorsImportance: warningDisk 1 - WDC_WD20EARS-00MVWB0_WD-WMAZA3638502 (sdh) (errors 1)Disk 3 - WDC_WD20EARS-00MVWB0_WD-WMAZA3690347 (sdj) (errors 52) syslog.txt
  17. FYI, i let the extended test finish and it passed. Attached is the smart test result for that drive. Was this just a fluke and I should rebuild the drive? nas-smart-20170301-1940.zip
  18. Thanks for your response. My machine was completely locked up and couldn't grab a diagnostics unfortunately. It was NOT a correcting parity check, just a default monthly. I am running an extended smart test right now on disk 4. It's been running for 4 hours now and at 50%. Should I let it finish or cancel, upgrade and run diagnostics again? Thanks
  19. I ran my monthly parity check last night and woke up to an unresponsive unraid. I got an alert "Disk 4 in error state" and another ""Array has 2 disks with read errors" around 1am. I rebooted, and was greeted with a red X over disk 4 "Disk is disabled, Contents emulated" I ran a smart test on disk4 with no errors. What are my next steps? Diagnostics attached. Thanks in advance. nas-diagnostics-20170301-0929.zip
  20. A new version of Radarr was just released with some nice improvements. Anyway we can get a midweek docker update?
  21. I've had my UPS working fine with Unraid for over 3 years. All of the sudden the battery charge is showing 0% and Runtime is 0 minutes. This is also causing my unraid to receive the shutdown command since I guess it thinks I have a power outage? My issue seems a lot like the one described here: https://lime-technology.com/forum/index.php?topic=45415.0 However a new usb port didnt fix it for me. Could my UPS be going bad? Output of /sbin/apcaccess: APC : 001,036,0870 DATE : 2017-02-01 18:55:21 -0600 HOSTNAME : nas VERSION : 3.14.13 (02 February 2015) slackware UPSNAME : nas CABLE : USB Cable DRIVER : USB UPS Driver UPSMODE : Stand Alone STARTTIME: 2017-02-01 18:55:19 -0600 MODEL : Back-UPS BX1000G STATUS : ONLINE LOWBATT LINEV : 117.0 Volts LOADPCT : 39.0 Percent BCHARGE : 0.0 Percent TIMELEFT : 0.0 Minutes MBATTCHG : 10 Percent MINTIMEL : 10 Minutes MAXTIME : 0 Seconds SENSE : Medium LOTRANS : 88.0 Volts HITRANS : 139.0 Volts ALARMDEL : No alarm BATTV : 10.4 Volts LASTXFER : No transfers since turnon NUMXFERS : 0 TONBATT : 0 Seconds CUMONBATT: 0 Seconds XOFFBATT : N/A SELFTEST : NO STATFLAG : 0x05000048 SERIALNO : 3B1138X10540 BATTDATE : 2011-09-14 NOMINV : 120 Volts NOMBATTV : 24.0 Volts NOMPOWER : 600 Watts FIRMWARE : 869.L2 .D USB FW:L2 END APC : 2017-02-01 18:55:37 -0600 root@nas:~#
  22. Yes, I know its you! And I really appreciate the assistance. I got it working with this config. Many thanks.
  23. Off the top of my head, I'd follow the instructions for nextcloud, but call the file requests.... Sent from my LG-H815 using Tapatalk Thanks, I did copy that config but was still greeted with the letsncrpyt landing page. Do I need to edit anything on the plex requests side? I saw in another post by CHBMB that you can edit the docker template but that seemed to be for /request instead of request.