Jump to content

Bait Fish

Members
  • Posts

    32
  • Joined

  • Last visited

Posts posted by Bait Fish

  1. Thanks for putting these up. I assume all of the blacklisting is either not needed or already done. My RPi install broke early this year and I never got back to it. Worked like a charm! [no, see update below]

     

    Using adsb-main of course, and adsb-fr24feed & adsb-flightaware. Can you put up a container for RadarBox too, please? Those are the three I was contributing to last. I'm still not sure about ads-b exchange after they got sold last year.

     

     

    Update: Actually, I failed to spot that adsb-flightaware is not happy actually. Trying to figure it out. The below keeps repeating (after my last try).

     

    Quote

    [piaware] 2024/06/08 12:13:24 lost connection to dump1090 via faup1090

    06/08/202412:13:25 PM

    [piaware] 2024/06/08 12:13:25 faup1090 exited with SIG SIGHUP

    06/08/202412:13:25 PM

    [piaware] 2024/06/08 12:13:25 will reconnect to dump1090 in 30 seconds

     

    I have to step away for other commitments now. Will continue later. Seems odd given fr24feed shows aircraft data sent, but not flightwaware.

     

     

    UPDATE 2, solved I think: Looks like I misinterpreted the "BEASTHOST" container setting. It needed an IP (192.168.x.x), not "Specify an external BEAST protocol provider (dump1090/readsb/etc)." I had readsb input there. Then it would not resolve the host. Rookie mistake. I now see data on flightware

     

  2. This is probably a no brainer for some, maybe not others. This app has been blacklisted in CA with no apparent updates, while the project page shows continuing update progress.

     

    I changed the container settings below and appear to have the latest version now.

     

    Repository:
    ghcr.io/swing-opensource/swingmusic:latest
    
    Registry URL:
    https://ghcr.io/swing-opensource/swingmusic:latest

     

  3. I had trouble viewing the GUI for the vnStat container. Defaults were left in the template during install except Network Type. I had set that to my custom network.

     

    The problem after first install was that clicking the WebUI link in Unraid would bring up a blank page using Chrome on my PC. This was the only machine/browser I tried. The URL listed was something link blocked/blank. Sorry. I did not capture the exact detail. This behaved differently from other containers. I tested using my machine's local IP and the container port. That too was not successful.

     

    I added this to the container and was able to reach the GUI. Click  Add another Path, Port, Variable, Label or Device and set the port details.

    Config Type: Port
    Name: Host Port for 8685
    Container Port: 8685
    Host Port: 8685
    Default Value: 8685
    Connection Type: TCP

     

    Once that was done it worked for me.

     

    Note that I had changed one other container before this success. I adjusted the WebUI setting (under advanced view). The default was "http://[IP]:8685/". I changed it to "http://[IP]:[PORT:8685]" to match other functioning containers on my system. I'm not sure if that had any bearing on getting the GUI to show.

     

    One last change was it had an inactive interface as far as I can tell. So after issuing these commands, my only active interface, eth0, was left. The two command issued start with vnstat.

     

    / # vnstat --iflist
    Available interfaces: tunl0 eth0 (10000 Mbit) 
    / # vnstat -i tunl0 --remove --force
    Interface "tunl0" removed from database.
    The interface will no longer be monitored. Use --add if monitoring the interface is again needed.
    / # 

     

    Hope this helps others.

  4. Worked for me like B4rny metioned. The 'Before you install' instructions listed in overview section of the container settings.

     

    One possible issue. The container defaults to appdata to /mnt/user/appdata/swing-music/ (hyphenated) which is different from the instructions ("cd swingmusic"). So, changed appdata to match where it built to.

  5. Uh oh. I have the VM autobackup itself on schedule. BUT I recall the morning the system went bad, that the VM backup executed.

     

    I may not.

     

    edit: libvirt.img is not in the backup location . . . I do not recall making any backup of it elsewhere manually.

     

    x2: and i did not have a location set in Appdata Backup/Restore for backing up libvirt.img

  6. root@Homer:/mnt/cache_nvme/system# ls -lah /mnt/cache_nvme/system/docker
    total 40G
    drwxrwxrwx 2 nobody users  24 Oct 11 16:32 ./
    drwxrwxrwx 4 nobody users  35 Nov 16  2021 ../
    -rw-rw-rw- 1 nobody users 50G Jan  1 01:39 docker.img
    root@Homer:/mnt/cache_nvme/system# ls -lah /mnt/cache_nvme/system/libvirt
    total 104M
    drwxrwxrwx 2 nobody users   25 Nov 16  2021 ./
    drwxrwxrwx 4 nobody users   35 Nov 16  2021 ../
    -rw-rw-rw- 1 nobody users 1.0G Dec 31 17:00 libvirt.img
    root@Homer:/mnt/cache_nvme/system# 

     

  7. I ran through a couple repair sessions and tried to follow its directions. I pasted all the notes in this code block including which flag I was using, typically keeping verbose on.

     

    I did not catch it telling me to run the L option so I have not done that. 

     

    Quote
    Phase 1 - find and verify superblock...
            - block cache size set to 3061336 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417567 tail block 393702
    ERROR: The filesystem has valuable metadata changes in a log which needs to
    be replayed.  Mount the filesystem to replay the log, and unmount it before
    re-running xfs_repair.  If you are unable to mount the filesystem, then use
    the -L option to destroy the log and attempt a repair.
    Note that destroying the log may cause corruption -- please attempt a mount
    of the filesystem before doing this.
    
    
    Stopped maint mode array. Started array. Stopped array. Started maint mode array. Rechecked xfs -nv
    
    Phase 1 - find and verify superblock...
            - block cache size set to 3061320 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417613 tail block 417613
            - scan filesystem freespace and inode maps...
            - found root inode chunk
    Phase 3 - for each AG...
            - scan (but don't clear) agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
    bad CRC for inode 3227792998
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
    No modify flag set, skipping phase 5
    Phase 6 - check inode connectivity...
            - traversing filesystem ...
            - agno = 0
            - agno = 1
    would rebuild directory inode 1239086149
            - agno = 2
            - agno = 3
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    disconnected inode 3227798067, would move to lost+found
    Phase 7 - verify link counts...
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117, can't compare link counts
    No modify flag set, skipping filesystem flush and exiting.
    
            XFS_REPAIR Summary    Tue Jan  3 16:05:11 2023
    
    Phase		Start		End		Duration
    Phase 1:	01/03 16:05:02	01/03 16:05:02
    Phase 2:	01/03 16:05:02	01/03 16:05:02
    Phase 3:	01/03 16:05:02	01/03 16:05:07	5 seconds
    Phase 4:	01/03 16:05:07	01/03 16:05:08	1 second
    Phase 5:	Skipped
    Phase 6:	01/03 16:05:08	01/03 16:05:11	3 seconds
    Phase 7:	01/03 16:05:11	01/03 16:05:11
    
    Total run time: 9 seconds
    
    
    
    Ran xfs repair -v
    
    Phase 1 - find and verify superblock...
            - block cache size set to 3061320 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417613 tail block 417613
            - scan filesystem freespace and inode maps...
            - found root inode chunk
    Phase 3 - for each AG...
            - scan and clear agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
    bad CRC for inode 3227792998
    bad CRC for inode 3227792998, will rewrite
    cleared inode 3227792998
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 3
            - agno = 2
            - agno = 1
    clearing reflink flag on inode 2149504425
    clearing reflink flag on inode 1073935201
    clearing reflink flag on inode 2149760794
    clearing reflink flag on inode 3223369736
    clearing reflink flag on inode 2193593298
    clearing reflink flag on inode 3270882671
    clearing reflink flag on inode 286726471
    clearing reflink flag on inode 286743730
    clearing reflink flag on inode 3307207359
    clearing reflink flag on inode 1148980934
    clearing reflink flag on inode 1148980936
    clearing reflink flag on inode 1148980938
    clearing reflink flag on inode 1148980939
    clearing reflink flag on inode 1148980940
    clearing reflink flag on inode 1148980941
    clearing reflink flag on inode 1148980942
    clearing reflink flag on inode 1151756820
    clearing reflink flag on inode 1151756823
    clearing reflink flag on inode 325929482
    clearing reflink flag on inode 325929522
    clearing reflink flag on inode 325929602
    clearing reflink flag on inode 325929606
    clearing reflink flag on inode 325929607
    clearing reflink flag on inode 325929608
    clearing reflink flag on inode 3376954786
    clearing reflink flag on inode 1232940612
    clearing reflink flag on inode 1232940615
    clearing reflink flag on inode 1232940616
    clearing reflink flag on inode 1232940617
    clearing reflink flag on inode 379736619
    clearing reflink flag on inode 380663459
    clearing reflink flag on inode 380663497
    clearing reflink flag on inode 380663502
    clearing reflink flag on inode 380663503
    clearing reflink flag on inode 380663510
    clearing reflink flag on inode 380663511
    clearing reflink flag on inode 380663520
    clearing reflink flag on inode 380663521
    clearing reflink flag on inode 380663522
    clearing reflink flag on inode 380663523
    clearing reflink flag on inode 380663524
    clearing reflink flag on inode 380663525
    clearing reflink flag on inode 380663540
    clearing reflink flag on inode 380663541
    clearing reflink flag on inode 3410317497
    clearing reflink flag on inode 3410343811
    clearing reflink flag on inode 3410343893
    clearing reflink flag on inode 3410343927
    clearing reflink flag on inode 3410649145
    Phase 5 - rebuild AG headers and trees...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - reset superblock...
    Phase 6 - check inode connectivity...
            - resetting contents of realtime bitmap and summary inodes
            - traversing filesystem ...
            - agno = 0
            - agno = 1
    rebuilding directory inode 1239086149
            - agno = 2
            - agno = 3
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    Phase 7 - verify and correct link counts...
    
            XFS_REPAIR Summary    Tue Jan  3 16:07:43 2023
    
    Phase		Start		End		Duration
    Phase 1:	01/03 16:07:33	01/03 16:07:33
    Phase 2:	01/03 16:07:33	01/03 16:07:33
    Phase 3:	01/03 16:07:33	01/03 16:07:39	6 seconds
    Phase 4:	01/03 16:07:39	01/03 16:07:40	1 second
    Phase 5:	01/03 16:07:40	01/03 16:07:40
    Phase 6:	01/03 16:07:40	01/03 16:07:43	3 seconds
    Phase 7:	01/03 16:07:43	01/03 16:07:43
    
    Total run time: 10 seconds
    done
    
    
    xfs repair -nv
    
    Phase 1 - find and verify superblock...
            - block cache size set to 3061320 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417613 tail block 417613
            - scan filesystem freespace and inode maps...
            - found root inode chunk
    Phase 3 - for each AG...
            - scan (but don't clear) agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 1
            - agno = 3
            - agno = 2
    No modify flag set, skipping phase 5
    Phase 6 - check inode connectivity...
            - traversing filesystem ...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    Phase 7 - verify link counts...
    No modify flag set, skipping filesystem flush and exiting.
    
            XFS_REPAIR Summary    Tue Jan  3 16:08:50 2023
    
    Phase		Start		End		Duration
    Phase 1:	01/03 16:08:41	01/03 16:08:41
    Phase 2:	01/03 16:08:41	01/03 16:08:41
    Phase 3:	01/03 16:08:41	01/03 16:08:46	5 seconds
    Phase 4:	01/03 16:08:46	01/03 16:08:47	1 second
    Phase 5:	Skipped
    Phase 6:	01/03 16:08:47	01/03 16:08:50	3 seconds
    Phase 7:	01/03 16:08:50	01/03 16:08:50
    
    Total run time: 9 seconds
    
    
    xfs repair -v
    
    Phase 1 - find and verify superblock...
            - block cache size set to 3061320 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417613 tail block 417613
            - scan filesystem freespace and inode maps...
            - found root inode chunk
    Phase 3 - for each AG...
            - scan and clear agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 3
            - agno = 1
            - agno = 2
    Phase 5 - rebuild AG headers and trees...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - reset superblock...
    Phase 6 - check inode connectivity...
            - resetting contents of realtime bitmap and summary inodes
            - traversing filesystem ...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    Phase 7 - verify and correct link counts...
    
            XFS_REPAIR Summary    Tue Jan  3 16:10:45 2023
    
    Phase		Start		End		Duration
    Phase 1:	01/03 16:10:35	01/03 16:10:35
    Phase 2:	01/03 16:10:35	01/03 16:10:35
    Phase 3:	01/03 16:10:35	01/03 16:10:40	5 seconds
    Phase 4:	01/03 16:10:40	01/03 16:10:41	1 second
    Phase 5:	01/03 16:10:41	01/03 16:10:42	1 second
    Phase 6:	01/03 16:10:42	01/03 16:10:45	3 seconds
    Phase 7:	01/03 16:10:45	01/03 16:10:45
    
    Total run time: 10 seconds
    done

     

     

  8. Having trouble on mobile modifying the code box above. Scanned with the -nv flag per docs. 

    Quote

     

     

    Phase 1 - find and verify superblock...
            - block cache size set to 3061336 entries
    Phase 2 - using internal log
            - zero log...
    zero_log: head block 417567 tail block 393702
    ALERT: The filesystem has valuable metadata changes in a log which is being
    ignored because the -n option was used.  Expect spurious inconsistencies
    which may be resolved by first mounting the filesystem to replay the log.
            - scan filesystem freespace and inode maps...
    block (0,50297084-50297193) multiply claimed by cnt space tree, state - 2
    block (0,48998093-48998203) multiply claimed by cnt space tree, state - 2
    block (0,50379227-50379336) multiply claimed by cnt space tree, state - 2
    block (0,49633120-49633230) multiply claimed by cnt space tree, state - 2
    agf_freeblks 64128684, counted 64133581 in ag 0
    agf_freeblks 97418327, counted 97438471 in ag 2
    sb_icount 4368768, counted 4433280
    sb_ifree 44699, counted 1263236
    sb_fdblocks 340342643, counted 354388165
            - found root inode chunk
    Phase 3 - for each AG...
            - scan (but don't clear) agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
    data fork in ino 58802146 claims free block 7583440
    data fork in ino 58802146 claims free block 7583505
    data fork in ino 59896646 claims free block 7574017
    data fork in ino 59896646 claims free block 7574079
            - agno = 1
            - agno = 2
    bad nblocks 10397115 for inode 2175513269, would reset to 10397118
    bad nextents 207685 for inode 2175513269, would reset to 207683
            - agno = 3
    bad CRC for inode 3227792998
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
    free space (0,48811889-48811997) only seen by one free space btree
    free space (0,50494325-50494435) only seen by one free space btree
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 2
            - agno = 3
            - agno = 1
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
    bad nblocks 10397115 for inode 2175513269, would reset to 10397118
    bad nextents 207685 for inode 2175513269, would reset to 207683
    No modify flag set, skipping phase 5
    Phase 6 - check inode connectivity...
            - traversing filesystem ...
            - agno = 0
            - agno = 1
    would rebuild directory inode 1239086149
            - agno = 2
            - agno = 3
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    disconnected inode 3227798067, would move to lost+found
    Phase 7 - verify link counts...
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117, can't compare link counts
    No modify flag set, skipping filesystem flush and exiting.
    
            XFS_REPAIR Summary    Tue Jan  3 15:25:14 2023
    
    Phase		Start		End		Duration
    Phase 1:	01/03 15:25:04	01/03 15:25:04
    Phase 2:	01/03 15:25:04	01/03 15:25:04
    Phase 3:	01/03 15:25:04	01/03 15:25:09	5 seconds
    Phase 4:	01/03 15:25:09	01/03 15:25:10	1 second
    Phase 5:	Skipped
    Phase 6:	01/03 15:25:10	01/03 15:25:14	4 seconds
    Phase 7:	01/03 15:25:14	01/03 15:25:14
    
    Total run time: 10 seconds

     

  9. I started it up a second time. The cache drive (sde) that showed missing this morning, now showed present and ready. Nothing was done but restarting 9 hours later. 

     

    I shut it down, reseated the cables for cache sde.

     

    Then started back up, and cache sde remained available.  I saved diagnostics from this session and have uploaded to this post as suggested above.

     

    I'll attempt repairs now.

     

    Update:

    Cache_nvme repair with default -n option results with,

    Phase 1 - find and verify superblock...
    Phase 2 - using internal log
            - zero log...
    ALERT: The filesystem has valuable metadata changes in a log which is being
    ignored because the -n option was used.  Expect spurious inconsistencies
    which may be resolved by first mounting the filesystem to replay the log.
            - scan filesystem freespace and inode maps...
    block (0,50297084-50297193) multiply claimed by cnt space tree, state - 2
    block (0,48998093-48998203) multiply claimed by cnt space tree, state - 2
    block (0,50379227-50379336) multiply claimed by cnt space tree, state - 2
    block (0,49633120-49633230) multiply claimed by cnt space tree, state - 2
    agf_freeblks 64128684, counted 64133581 in ag 0
    agf_freeblks 97418327, counted 97438471 in ag 2
    sb_icount 4368768, counted 4433280
    sb_ifree 44699, counted 1263236
    sb_fdblocks 340342643, counted 354388165
            - found root inode chunk
    Phase 3 - for each AG...
            - scan (but don't clear) agi unlinked lists...
            - process known inodes and perform inode discovery...
            - agno = 0
    data fork in ino 58802146 claims free block 7583440
    data fork in ino 58802146 claims free block 7583505
    data fork in ino 59896646 claims free block 7574017
    data fork in ino 59896646 claims free block 7574079
            - agno = 1
            - agno = 2
    bad nblocks 10397115 for inode 2175513269, would reset to 10397118
    bad nextents 207685 for inode 2175513269, would reset to 207683
            - agno = 3
    bad CRC for inode 3227792998
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
            - process newly discovered inodes...
    Phase 4 - check for duplicate blocks...
            - setting up duplicate extent list...
    free space (0,48811889-48811997) only seen by one free space btree
    free space (0,50494325-50494435) only seen by one free space btree
            - check for inodes claiming duplicate blocks...
            - agno = 0
            - agno = 1
            - agno = 2
            - agno = 3
    bad CRC for inode 3227792998, would rewrite
    would have cleared inode 3227792998
    bad nblocks 10397115 for inode 2175513269, would reset to 10397118
    bad nextents 207685 for inode 2175513269, would reset to 207683
    No modify flag set, skipping phase 5
    Phase 6 - check inode connectivity...
            - traversing filesystem ...
    would rebuild directory inode 1239086149
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117
            - traversal finished ...
            - moving disconnected inodes to lost+found ...
    disconnected inode 3227798067, would move to lost+found
    Phase 7 - verify link counts...
    Metadata corruption detected at 0x46e010, inode 0xc0643666 dinode
    couldn't map inode 3227792998, err = 117, can't compare link counts
    No modify flag set, skipping filesystem flush and exiting.

     

    homer-diagnostics-20230103-1458.zip

  10. Now that I'm back in the land of the living, details I could not include in my haste this morning. These errors happened in the middle of the night on a system that was humming along without any other concerns previously.

     

    Shares are not showing that had been there previously. I'm sure this is a good part of the problem.

     

    Unraid Version: 6.11.5. 

     

    Fix Common Problems greeted me with these issues when I checked on the server after coffee:

    Unable to write to cache_nvme    Drive mounted read-only or completely full.
    Unable to write to cache_ssd    Drive mounted read-only or completely full. 
    Unable to write to Docker Image    Docker Image either full or corrupted.

     

    On the Docker tab in Unraid, these errors show.

    Docker Containers
    APPLICATION    VERSION    NETWORK    PORT MAPPINGS (APP TO HOST)    VOLUME MAPPINGS (APP TO HOST)    AUTOSTART    UPTIME    

    Warning: stream_socket_client(): unable to connect to unix:///var/run/docker.sock (Connection refused) in /usr/local/emhttp/plugins/dynamix.docker.manager/include/DockerClient.php on line 712
    Couldn't create socket: [111] Connection refused
    Warning: Invalid argument supplied for foreach() in /usr/local/emhttp/plugins/dynamix.docker.manager/include/DockerClient.php on line 898

    Warning: stream_socket_client(): unable to connect to unix:///var/run/docker.sock (Connection refused) in /usr/local/emhttp/plugins/dynamix.docker.manager/include/DockerClient.php on line 712
    Couldn't create socket: [111] Connection refused
    Warning: Invalid argument supplied for foreach() in /usr/local/emhttp/plugins/dynamix.docker.manager/include/DockerClient.php on line 967
    No Docker containers installed

     

    And Unraid Settings>Docker mirrors these errors.

    Enable Docker:

     Yes  One or more paths do not exist (view)

    Docker vDisk location:

      /mnt/user/system/docker/docker.img  Path does not exist

    Default appdata storage location:

      /mnt/user/appdata/  Path does not exist

  11. Following up. I did not get a good extended test log via Unraid. I instead swapped the suspect drive out and got its replacement going first. Attempts at testing it as an external drive kept failing, but without log as far as I could tell. Scanning it with Seatools on Windows ended with a failure in the long test warning that the drive is... failing. Thanks again, all of you, for your help. Unraid, and its communitity, are awesome!

     

    --------------- SeaTools for Windows v1.4.0.7 ---------------
    5/24/2019 3:17:39 PM
    Model Number: Seagate Backup+ Hub BK
    Serial Number: NA8TGN5Z
    Firmware Revision: D781
    SMART - Started 5/24/2019 3:17:39 PM
    SMART - Pass 5/24/2019 3:17:45 PM
    Short DST - Started 5/24/2019 3:17:52 PM
    Short DST - Pass 5/24/2019 3:18:58 PM
    Identify - Started 5/24/2019 3:19:03 PM
    Short Generic - Started 5/24/2019 3:25:19 PM
    Short Generic - Pass 5/24/2019 3:26:31 PM
    Identify - Started 5/3/2022 4:36:31 PM
    Short DST - Started 5/3/2022 4:36:46 PM
    Short DST - Pass 5/3/2022 4:37:58 PM
    Short Generic - Started 5/3/2022 4:38:45 PM
    Short Generic - Pass 5/3/2022 4:40:27 PM
    Long Generic - Started 5/3/2022 4:41:42 PM
    Long Generic - FAIL 5/4/2022 1:18:40 AM
    SeaTools Test Code: E896A6D4

     

  12. Thanks for the tips and insights. Making progress now that spin down is disabled. So simple. . . I'll remember this next time.

     

    Your cautions spurred me to quit new disk intensive activity I started today. Now most everything is stopped, disk activity at a minimum.

     

    The new replacement 8TB drive's preclear is going to finish soon, 2% post-read left. To get the array to normal sooner, I'll play it safe by adding the new drive in first, rebuilding, then testing the failing drive later while unassigned.

     

     

×
×
  • Create New...