jimbo123

Members
  • Posts

    44
  • Joined

  • Last visited

Posts posted by jimbo123

  1. Hardware is one of the early HP MIcroservers,.. an N36L. Nothing changed really, running onboard controller. I updated BIOS firmware to enable the CD/DVD drive to operate as a SATA disk, or something similar from memory.

     

    One parity disk and 4 data drives as mentioned earlier.

     

    I'll update to current stable and see how things behave.

  2. Hi all,

     

    I've been running unRAID for about 10 years now,.. mostly on auto-pilot, with only issues I've had a a few failed drives.

     

    I run a monthly parity check,.. and up until the last few weeks have never had any sync errors with the check at all.

     

    The last few parity checks have been showing up corrected errors.

     

    Today 1st Nov,.. it did its normal monthly check and again found and corrected sync errors (217).

    Nov  1 00:00:01 storage kernel: mdcmd (42): check CORRECT
    Nov  1 00:00:01 storage kernel: 
    Nov  1 00:00:01 storage kernel: md: recovery thread woken up ...
    Nov  1 00:00:01 storage kernel: md: recovery thread checking parity...
    Nov  1 00:00:01 storage kernel: md: using 1536k window, over a total of 3907018532 blocks.
    Nov  1 00:05:16 storage kernel: md: correcting parity, sector=65721672
    Nov  1 00:09:27 storage kernel: md: correcting parity, sector=122070112
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122183264
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122187616
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122196320
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122213728
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122218080
    Nov  1 00:09:28 storage kernel: md: correcting parity, sector=122252896
    Nov  1 00:09:29 storage kernel: md: correcting parity, sector=122435680
    Nov  1 00:09:29 storage kernel: md: correcting parity, sector=122440032
    Nov  1 00:09:29 storage kernel: md: correcting parity, sector=122514016
    Nov  1 00:09:29 storage kernel: md: correcting parity, sector=122540128
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122548832
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122557536
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122561888
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122566240
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122592352
    Nov  1 00:09:30 storage kernel: md: correcting parity, sector=122596704
    Nov  1 00:20:28 storage kernel: md: correcting parity, sector=270032168
     

    After it had finished,.. I manually kicked off the check again after a few hours,.. and again it found errors again, (its almost done and is at 218).

    Nov  1 14:15:11 storage kernel: md: recovery thread woken up ...
    Nov  1 14:15:11 storage kernel: md: recovery thread checking parity...
    Nov  1 14:15:11 storage kernel: md: using 1536k window, over a total of 3907018532 blocks.
    Nov  1 14:20:20 storage kernel: md: correcting parity, sector=65721672
    Nov  1 14:24:08 storage kernel: md: correcting parity, sector=117798280
    Nov  1 14:24:13 storage kernel: md: correcting parity, sector=118812296
    Nov  1 14:24:15 storage kernel: md: correcting parity, sector=119217032
    Nov  1 14:24:15 storage kernel: md: correcting parity, sector=119260552
    Nov  1 14:24:16 storage kernel: md: correcting parity, sector=119452040
    Nov  1 14:24:27 storage kernel: md: correcting parity, sector=122070112
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122183264
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122187616
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122196320
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122213728
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122218080
    Nov  1 14:24:28 storage kernel: md: correcting parity, sector=122252896
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122435680
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122440032
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122514016
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122540128
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122548832
    Nov  1 14:24:29 storage kernel: md: correcting parity, sector=122557536

     

    Have I got a disk that is on its way out ? I see no other errors in the log,.. previously, when I was having disk errors, there'd be something in there.

     

    I'm running 5.0.5 with the following layout:
    4TB parity
    2 x 2TB
    2 x 4TB

     

    I've looked at smartctl details and the only thing that looks out of place is in the attached file (WDC_WD40EFRX-68N32N0_3704261.txt), but the events in there seem like they're old, but I'm no expert here. The other 4 drives show similar to the file 'WDC_WD40EFRX-68N32N0_WD-WCC7K5NPLNN1.txt'

     

    Can the sectors associated with the errors be matched back to a disk ?

     

    Hoping someone can shed some light on what's going on and help me identify if I do have a disk that needs to be replaced.

     

    Thanks in advance,

     

    Jim.....

    WDC_WD40EFRX-68N32N0_3704261.txt WDC_WD40EFRX-68N32N0_WD-WCC7K5NPLNN1.txt

  3. Drive arrived during the week,.. new drive is now going through a preclear,.. expect it should finish overnight and I can then start the configuring when I get home from work tomorrow.

     

    So the situation I'm in is:

    • Went through steps 1 to 8
    • Array is running minus the the drive I pulled.
    • New drive is preclearing in its slot.

    I assume that when the preclear finishes, that I:

    8b. stop the array

    9. unassign the parity disk (can I do this without losing any data ? as the old disk is not replaced yet and I'll now also be a parity drive out ?)

    10. assign the new drive as the parity

    11. assign old parity drive into the array

    12, 13 , 14 as per guide.

     

    Step 9 is what I'm unsure of ? 

     

    Or should I add:

    8c. assign the new drive into the array (taking the slot of the failing drive I removed) - let it rebuild and sort itself out

     

    and then continue with step 9 and onwards ?

     

    Jim.....

  4. Hi all,

     

    I currently have an unRAID v5.0.5 system that's been running on virtually auto-pilot for several years. It runs a monthly parity check and has always indicated no issue. The logs below occurred during this months parity check and it still showed up with 0 parity errors.

     

    Setup is:

    • Hardware: HP ProLiant MicroServer N36L 
    • Data: 4 x 2TB WD20EARS
    • Parity: 2TB WD20EARS

     

    Have recently noticed the following:

    Nov  1 05:14:42 storage kernel: ata1.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0 (Errors)
    Nov  1 05:14:42 storage kernel: ata1.00: irq_stat 0x40000001 (Drive related)
    Nov  1 05:14:42 storage kernel: ata1.00: failed command: READ DMA EXT (Minor Issues)
    Nov  1 05:14:42 storage kernel: ata1.00: cmd 25/00:00:68:62:72/00:04:b7:00:00/e0 tag 0 dma 524288 in (Drive related)
    Nov  1 05:14:42 storage kernel:          res 51/40:9f:c8:64:72/00:01:b7:00:00/e0 Emask 0x9 (media error) (Errors)
    Nov  1 05:14:42 storage kernel: ata1.00: status: { DRDY ERR } (Drive related)
    Nov  1 05:14:42 storage kernel: ata1.00: error: { UNC } (Errors)
    Nov  1 05:14:42 storage kernel: ata1.00: configured for UDMA/133 (Drive related)
    Nov  1 05:14:42 storage kernel: sd 1:0:0:0: [sdb] Unhandled sense code (Drive related)
    Nov  1 05:14:42 storage kernel: sd 1:0:0:0: [sdb]   (Drive related)
    Nov  1 05:14:42 storage kernel: Result: hostbyte=0x00 driverbyte=0x08 (System)
    Nov  1 05:14:42 storage kernel: sd 1:0:0:0: [sdb]   (Drive related)
    Nov  1 05:14:42 storage kernel: Sense Key : 0x3 [current] [descriptor]
    Nov  1 05:14:42 storage kernel: Descriptor sense data with sense descriptors (in hex):
    Nov  1 05:14:42 storage kernel:         72 03 11 04 00 00 00 0c 00 0a 80 00 00 00 00 00 
    Nov  1 05:14:42 storage kernel:         b7 72 64 c8 
    Nov  1 05:14:42 storage kernel: sd 1:0:0:0: [sdb]   (Drive related)
    Nov  1 05:14:42 storage kernel: ASC=0x11 ASCQ=0x4
    Nov  1 05:14:42 storage kernel: sd 1:0:0:0: [sdb] CDB:  (Drive related)
    Nov  1 05:14:42 storage kernel: cdb[0]=0x28: 28 00 b7 72 62 68 00 04 00 00
    Nov  1 05:14:42 storage kernel: end_request: I/O error, dev sdb, sector 3077727432 (Errors)
    Nov  1 05:14:42 storage kernel: ata1: EH complete (Drive related)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727368 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727376 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727384 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727392 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727400 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727408 (Errors)
    Nov  1 05:14:42 storage kernel: md: disk1 read error, sector=3077727416 (Errors)

     

    Disk 1 has about 1TB of data on it. All up the array is about 50% utilised.

     

    I figure the disk above is on its way out, so a WD40EFRX 4TB RED drive is on its way.

     

    What I'd like to do is:

    • make the new 4TB the parity disk
    • have the existing parity assume the role of the failing disk 1

     

    Have found this link,.. the block titled "The procedure" seems relevent to me.
    https://wiki.lime-technology.com/The_parity_swap_procedure

     

    Every guide I've read always recommends preclearing, but have also read that its not necessary if its a parity drive ?

     

    I'm figuring the following is what i need to do:

    1. Given I have no spare SATA spots in my system, I follow steps 1 to 8.
    2. New drive will then be unassigned.
    3. Run the pre-clear script against the new drive ? If its needed ?
    4. Continue on from step 9

     

    So my queries:

    • is preclear needed if its role will be that of a parity drive ?
    • am I following the right guide/steps for what I want to do ?
    • am I missing anything ?

     

    Any guidance would be appreciated.

     

    Thanks in advance,

     

    Jim.....

     

    smart.txt

  5. I'm in a similar situation,... the approach I was going to take was to connect my UPS to a server that supported the USB connection and run upsd on there along with upsmon and then from unRAID's perspective attempt at getting upsmon happening.

     

    Haven't given it much thought aside of the above and thought it wasn't going to be too difficult ??

     

    I'll find out in the next few weeks.