justintas

Members
  • Posts

    19
  • Joined

  • Last visited

Everything posted by justintas

  1. Hi, did my regular Parity check and noticed following error, any advice appreciated Thanks Justintas Jun 30 19:35:31 HPtower kernel: BTRFS error (device loop3): parent transid verify failed on 30883840 wanted 347 found 18 Jun 30 19:35:31 HPtower root: mount: /etc/libvirt: wrong fs type, bad option, bad superblock on /dev/loop3, missing codepage or helper program, or other error. Jun 30 19:35:31 HPtower root: mount error Jun 30 19:35:31 HPtower kernel: BTRFS error (device loop3): parent transid verify failed on 30883840 wanted 347 found 18 Jun 30 19:35:31 HPtower kernel: BTRFS warning (device loop3): couldn't read tree root Jun 30 19:35:31 HPtower kernel: BTRFS error (device loop3): open_ctree failed hptower-diagnostics-20220701-1844.zip
  2. Should I do another parity rebuild before changing parity disk to larger disk ?
  3. Ok tried check process again here are the results below, and guess what it is fixed !!! Thanks for your guidance much appreciated re ran check -n Results xfs_repair status: Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... - scan filesystem freespace and inode maps... - found root inode chunk Phase 3 - for each AG... - scan (but don't clear) agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 0 - agno = 1 - agno = 2 - agno = 3 No modify flag set, skipping phase 5 Phase 6 - check inode connectivity... - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify link counts... would have reset inode 4328816385 nlinks from 1 to 2 would have reset inode 4328816389 nlinks from 1 to 2 would have reset inode 4328816398 nlinks from 1 to 2 No modify flag set, skipping filesystem flush and exiting. Then re ran check (blank) to do a repair Results: Phase 1 - find and verify superblock... Phase 2 - using internal log - zero log... - scan filesystem freespace and inode maps... - found root inode chunk Phase 3 - for each AG... - scan and clear agi unlinked lists... - process known inodes and perform inode discovery... - agno = 0 - agno = 1 - agno = 2 - agno = 3 - process newly discovered inodes... Phase 4 - check for duplicate blocks... - setting up duplicate extent list... - check for inodes claiming duplicate blocks... - agno = 0 - agno = 1 - agno = 2 - agno = 3 Phase 5 - rebuild AG headers and trees... - reset superblock... Phase 6 - check inode connectivity... - resetting contents of realtime bitmap and summary inodes - traversing filesystem ... - traversal finished ... - moving disconnected inodes to lost+found ... Phase 7 - verify and correct link counts... resetting inode 4328816385 nlinks from 1 to 2 resetting inode 4328816389 nlinks from 1 to 2 resetting inode 4328816398 nlinks from 1 to 2 done
  4. Yes its jammed in the cage and cant be removed screw must have moved yes is disk 1 is green but showing as unmountable , mounting existing drive will be hard as cage damaged. Will see if can cut drive out of cage any other options or do pictures attached if any help
  5. check and repair was from gui check first with default settings -n had to put array into mainteance mode to run then ran check again with a blank option was that the correct steps ?
  6. yes original disk is jammed in disk cage hence swapped to new cage and put new disk in as disk 1 (sdc) Are errors recoverable ?
  7. Ok ran check then repair output doesn't look to good ? Here is last lines of process.. Metadata corruption detected at 0x44d778, xfs_bmbt block 0xec37d798/0x1000 libxfs_bwrite: write verifier failed on xfs_bmbt bno 0xec37d798/0x1000 Maximum metadata LSN (2146145896:-2144772351) is ahead of log (22:71999). Format log to cycle 2146145899. xfs_repair: Releasing dirty buffer to free list! cache_purge: shake on cache 0x5021c0 left 3 nodes!? xfs_repair: Refusing to write a corrupt buffer to the data device! xfs_repair: Lost a write to the data device! fatal error -- File system metadata writeout failed, err=117. Re-run xfs_repair. Options ? above run in mainteance mode only way to activate check option I do have another disk available to try as a replacement for SDC ? Original SDC is still available but jammed in drive cage
  8. just struck a problem with fixing above had a delay due to a damaged cage drive so had to switch hardware. Have inserted new disk 1 and did a rebuild. When reboot 2 things are happening; 1. it is saying disk 1 ''Unmountable disk present: '' not sure what did wrong here ? 2. All my dockers are not visible assume could be related to 1 above. updated diagnostics attached , have tried 2 rebuilds but can't work out what i have done wrong ? hptower-diagnostics-20211215-1157.zip
  9. Thanks Trurl , will wait to current parity option finishes, about an hour then follow the steps you have outlined. Assume once data disk 1 is rebuilt and all boots ok can go ahead again and replace the parity disk ? then start gradual upgrade of each data disk. Really appreciate your help and advice.
  10. ok , so I have a replacement ready brand new but is same size as original parity 8tb is that ok ? or an older 4tb drive which one to use? So what steps do I follow , assume let parity rebuild follow first
  11. Thanks for your help to answer questions; No notification about disk 1, Yes it has a notification on dashboard now about error, No wasn't reason decided to replace parity I need to expand array as running out of disk so replacing parity first. No nothing written to array since rebuild started, yes still have original parity disk, yes copies of most of important data etc , just my movie collection on there that is not backed up. Options ?
  12. Diagnostics attached hptower-diagnostics-20211130-1009.zip
  13. Hoping for some direction on error encountered. Array running out of space so started process to replace Parity Disk with a larger one, all going ok till struck an error in rebuild it is a read error of of one of the disks see below. Current rebuild has got another 3 hours to go to finish. Prior to parity upgrade was no errors and last parity rebuild was all good. Question , Is this something to worry about and will it lead to some data loss ? When it finishes if reboot will the error correct itself ? I have new data disks to install and old parity disk still available. Any help or suggestions appreciated Thanks in advance Justintas Parity - ST12000VN0008-2PH103_ZS802V5R (sdb) - active 33 C [DISK INVALID] (new parity Disk getting rebuilt) Disk 1 - WDC_WD40EFRX-68N32N0_WD-WCC7K2PF6VZX (sdc) - active 31 C (disk has read errors) [NOK] Disk 2 - WDC_WD40EFRX-68N32N0_WD-WCC7K3EN931N (sdd) - active 31 C [OK] Disk 3 - WDC_WD40EFRX-68N32N0_WD-WCC7K5FREU23 (sde) - active 31 C [OK] Parity sync / Data rebuild in progress. Total size: 12 TB Elapsed time: 8 hours, 12 minutes Current position: 4.01 TB (33.4 %) Estimated speed: 216.3 MB/sec Estimated finish: 10 hours, 16 minutes Sync errors corrected: 2689
  14. Thanks @itimpi looks like it is my pihole docker will fix reference to cache and try again
  15. Hi, hope you can point me in the right direction to solve this error. Updated Unraid and also Fix Common problem tool and have had a new error appear hav'nt changed anything else not sure where to look or if I can ignore. Log file attached appreciate your help. Cheers Justintas hptower-syslog-20200419-0838.zip hptower-diagnostics-20200419-1914.zip
  16. I like how easy it was to setup and use with the GUI interface I would like better integration into a windows AD environment for the OS in 2020, as we use Unraid as a network backup for our server / network