Jump to content

CyberMew

Members
  • Posts

    410
  • Joined

Posts posted by CyberMew

  1. Hi. I am getting some errors in log, not sure why, I think my docker containers all stopped due to backup (not sure), but it is not resuming and I don't see the weekly backup that it usually runs.

    Quote

    Jan 9 04:55:41 Tower kernel: BTRFS info (device loop4): no csum found for inode 3887 start 237568
    Jan 9 04:55:41 Tower kernel: BTRFS warning (device loop4): csum failed root 396 ino 3887 off 237568 csum 0x7f0a4dc5 expected csum 0x00000000 mirror 1
    Jan 9 04:55:41 Tower kernel: BTRFS error (device loop4): bdev /dev/loop4 errs: wr 0, rd 0, flush 0, corrupt 46726, gen 0
    Jan 9 04:55:41 Tower kernel: BTRFS error (device loop4): tree first key mismatch detected, bytenr=196624384 parent_transid=235 key expected=(18446744073709535222,128,5549367296) has=(18446744073709551606,128,5549367296)
    Jan 9 04:55:41 Tower kernel: BTRFS error (device loop4): tree first key mismatch detected, bytenr=196624384 parent_transid=235 key expected=(18446744073709535222,128,5549367296) has=(18446744073709551606,128,5549367296)

    Any idea what is the issue and how to fix it?

    tower-diagnostics-20240110-2303.zip

  2. 1 minute ago, itimpi said:

    I don’t think the recommendation has changed 😊. As to whether the default setting has changed I am not sure :😒

    Oh hmm I don’t think I changed something like this.. I did remember someone saying it’s supposed to write if during parity check there were errors, since there’s no way it can be fixed anyway..

    How can I verify my setting and change it back to default?

  3. So there are no unexpected shutdowns. The drive replacement was done successfully. And one month later this thing happened. Not sure why? Either parity drive was corrupted or some other drives were corrupted?


    I will try the link you provided and report back. 
     

    As for the parity correcting thing you mentioned, I think it is the default setup I’ve been since using since 5.x. Did something change? 

     

  4. 24 minutes ago, Can0n said:



    found it!! thank you for your help

    root@Thor:~# nano /etc/cron.d/root
    grep -r "Backup Share to Freya" /boot/
    /boot/config/plugins/dynamix/thor2freya.cron:#*** rsync Thor's Backup Share to Freya

    Thanks so much for sharing your findings! Mine was placed there as well. I think this was the post I came across last time

     Finally, I can resolve it now. 

    • Like 1
  5. No, it doesn't get saved for me. I have a similar issue and I have no idea where I added that cron job, and no matter how I modify the /etc/cron.d/root file, it always revert back after I restart the server or start the array. I took some instructions on this forum very long ago as there was no such thing as user scripts back then and have no idea how to reset/remove it now

  6. On 6/6/2023 at 12:24 AM, JorgeB said:

    So it's solved?

    Yes seems like it, thanks a lot! I guess after copy and rebuild operation there is a startup bug and it didn't try to attempt to start the services.. (I hope)

     

    Disk1 scan was also ok:

    Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error

    # 1 Extended offline Completed without error 00% 64443 -

  7. 5 hours ago, JorgeB said:

    I don't see any attempt to start the docker service, try disabling the service apply, re-enable, apply and post new diags.

     

    P.S. you should run an extended SMART test on disk1, appears to be failing.

    I googled and found a similar issue 

    I restarted and that seemed to do the trick! Attached new diags just in case.tower-diagnostics-20230605-2330.zip

     

    I am also running the extended SMART test now. Thanks a lot!

     

    5 hours ago, itimpi said:

    You are getting the following error repeating in the syslog:

    Jun  4 10:48:01 Tower crond[1974]: failed parsing crontab for user root: #015

    You should perhaps post the output of 

    cat /etc/crond.d/root

    if it is not obvious what is causing that.

     

    You also have the following:

    Jun  5 02:33:34 Tower kernel: sd 10:0:3:0: [sdi] tag#1210 UNKNOWN(0x2003) Result: hostbyte=0x00 driverbyte=0x08 cmd_age=6s
    Jun  5 02:33:34 Tower kernel: sd 10:0:3:0: [sdi] tag#1210 Sense Key : 0x3 [current] [descriptor] 
    Jun  5 02:33:34 Tower kernel: sd 10:0:3:0: [sdi] tag#1210 ASC=0x11 ASCQ=0x0 
    Jun  5 02:33:34 Tower kernel: sd 10:0:3:0: [sdi] tag#1210 CDB: opcode=0x88 88 00 00 00 00 02 8f 36 16 20 00 00 04 00 00 00
    Jun  5 02:33:34 Tower kernel: blk_update_request: critical medium error, dev sdi, sector 10992621088 op 0x0:(READ) flags 0x4000 phys_seg 128 prio class 0
    Jun  5 02:33:34 Tower kernel: md: disk1 read error, sector=10992621024
    Jun  5 02:33:34 Tower kernel: md: disk1 read error, sector=10992621032
    Jun  5 02:33:34 Tower kernel: md: disk1 read error, sector=10992621040

    which looks like it may be a genuine disk issue so you should consider running an extended SMART test on it as a check.

     

    There are also a lot of FCP warnings that it might be a good idea to consider tidying up.

     

     

     

    Thanks for the suggestions! Yes I should definitely get around to FCP..

     

    The command "cat /etc/crond.d/root" is not working but I think you meant this:
     

    ~# cat /etc/cron.d/root

    # Generated docker monitoring schedule:

    10 0 * * 1 /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/dockerupdate.php check &> /dev/null

     

    # Generated system monitoring schedule:

    */1 * * * * /usr/local/emhttp/plugins/dynamix/scripts/monitor &> /dev/null

     

    # Generated mover schedule:

    40 4 * * * /usr/local/sbin/mover &> /dev/null

     

    # Generated parity check schedule:

    0 0 1 * * /usr/local/sbin/mdcmd check  &> /dev/null

     

    # Generated plugins version check schedule:

    10 0 * * * /usr/local/emhttp/plugins/dynamix.plugin.manager/scripts/plugincheck &> /dev/null

     

    #Refresh plex tv shows at 3pm everyday 0 15

    36 20 * * * logger -tscriptRefreshPlexLibrary5[$$] "Refreshing Plex library 5" && curl "http://127.0.0.1:32400/library/sections/5/refresh?deep=0&X-Plex-Token=xxx" &> /dev/null <- any idea how to remove this?

     

    # Generated array status check schedule:

    20 0 * * * /usr/local/emhttp/plugins/dynamix/scripts/statuscheck &> /dev/null

     

    #Running subliminal At minute 0 past every 4th hour. cat /etc/cron.d/root

    0 */4 * * * logger -tscriptSubliminal[$$] "Subliminal checking for subs" && /mnt/user/subliminal2/subliminal/checklast2dayspath.sh > /dev/null <- any idea how to remove this as well?

     

    # Generated cron settings for docker autoupdates

    0 0 * * 0 /usr/local/emhttp/plugins/ca.update.applications/scripts/updateDocker.php >/dev/null 2>&1

    # Generated cron settings for plugin autoupdates

    0 0 * * * /usr/local/emhttp/plugins/ca.update.applications/scripts/updateApplications.php >/dev/null 2>&1

     

    # CRON for CA background scanning of applications

    34 * * * * php /usr/local/emhttp/plugins/community.applications/scripts/notices.php > /dev/null 2>&1

     

    # Generated ssd trim schedule:

    0 3 * * * /sbin/fstrim -a -v | logger &> /dev/null

     

    Not sure what is # 015..

  8. On 6/2/2023 at 6:52 PM, JorgeB said:

    Disk itself looks OK but it's generating this error:

     

    Jun  2 06:38:09 Tower kernel: md: import disk0: lock_bdev error: -13

     

    which I don't remember seeing before, try swapping that disk with one using the onboard SATA controller to rule out some compatibility issue with the HBA.

    Thanks! I switch it and it seems to be working fine as expected. Copied without errors over the past day and rebuilding disk5 now. 
     

    Edit: is the add on card problematic? Should I change it? Any brands or model to recommend?

  9. I got around to doing this today, and for some reason Parity 1 slot is not showing the new drive, but Parity 2 slot is able to show/select it. I want to use the bigger HDD for Parity 1 instead. Is this not possible?

     

    Current:

    2100361378_CleanShot2023-06-02at06_41.09@2x.png.e5ef7afd16a22790aa16d6be47225ea4.png

     

    Edit: after moving the drives around the new drive is somehow appearing under Parity 1 to be selectable. Not sure if this was a bug in 6.9.2.

     

    Edit2: After selecting the new drive in Parity 1 slot, it does not show up at all! Is the wiki instructions outdated?1305081895_CleanShot2023-06-02at06_45.17@2x.thumb.png.41ce88e622555b94b32196789a1f65e9.png

  10. I tried to preclear using the binhex docker as suggested, and it precleared successfully. 1955386023_CleanShot2023-06-01at11_16.33@2x.png.0b641911272eae44a1df776259db459f.png

     

    So maybe there is some bug with the plugin..? Should I stop preclear and Verify Disk using this plugin, or is the drive good to go now?

     

    Quote

    ================================================================== 1.22
    =                unRAID server Pre-Clear disk /dev/sdk
    =               cycle 1 of 1, partition start on sector 64
    = Disk Pre-Clear-Read completed                                 DONE
    = Step 1 of 10 - Copying zeros to first 2048k bytes             DONE
    = Step 2 of 10 - Copying zeros to remainder of disk to clear it DONE
    = Step 3 of 10 - Disk is now cleared from MBR onward.           DONE
    = Step 4 of 10 - Clearing MBR bytes for partition 2,3 & 4       DONE
    = Step 5 of 10 - Clearing MBR code area                         DONE
    = Step 6 of 10 - Setting MBR signature bytes                    DONE
    = Step 7 of 10 - Setting partition 1 to precleared state        DONE
    = Step 8 of 10 - Notifying kernel we changed the partitioning   DONE
    = Step 9 of 10 - Creating the /dev/disk/by* entries             DONE
    = Step 10 of 10 - Verifying if the MBR is cleared.              DONE
    = Disk Post-Clear-Read completed                                DONE
    Disk Temperature: 56C, Elapsed Time:  101:02:38
    ========================================================================1.22
    == WDCWD181KFGX-68AFPN0  
    == Disk /dev/sdk has been successfully precleared
    == with a starting sector of 64
    ============================================================================
    ** Changed attributes in files: /tmp/smart_start_sdk  /tmp/smart_finish_sdk
                    ATTRIBUTE   NEW_VAL OLD_VAL FAILURE_THRESHOLD STATUS      RAW_VALUE
          Temperature_Celsius =    35      53            0        ok          56
     No SMART attributes are FAILING_NOW

     0 sectors were pending re-allocation before the start of the preclear.
     0 sectors were pending re-allocation after pre-read in cycle 1 of 1.
     0 sectors were pending re-allocation after zero of disk in cycle 1 of 1.
     0 sectors are pending re-allocation at the end of the preclear,
        the number of sectors pending re-allocation did not change.
     0 sectors had been re-allocated before the start of the preclear.
     0 sectors are re-allocated at the end of the preclear,
        the number of sectors re-allocated did not change.

     

  11. 7 hours ago, dlandon said:

    That drive should be emulated, assuming you have parity, and you can start the array.

     

    Use 'diagnostics' on a command line to get diagnottics.  Preclear log can be downloaded from the UD Preclear page.  Click on the download icon.

    My disk 5 is unassigned right now with a red X, will it know to emulate disk 5 or remove it and update my parity (losing data)?

     

    Also have uploaded the requested logs here:

    TOWER-unassigned.devices.preclear-20230527-0806.zip tower-diagnostics-20230527-0048.zip

×
×
  • Create New...