Jump to content

-Daedalus

Members
  • Posts

    426
  • Joined

  • Last visited

Posts posted by -Daedalus

  1. In order of preference:

     

    Native backup solution (for pools and array to other pool/array, unRAID server, or network location). There's a plug-in, I know, but something more granular than a scheduled zip would be nice.

     

    VM backup and snapshots from GUI

     

    A way to remove a disk (array shrink) without putting data at risk (at no point should parity be rebuilding without redundancy, or invalid).

     

    • Upvote 6
  2. Thanks for the quick response!

     

    They were overheating originally, so I added some heatsinks, though my understanding is that the SSDs with throttle performance to avoid errors just like this, so I'm not quite sure what's going on.

  3. Google-fu not great here. BTRFS and what look like errors from the controller from the drive (Samsung 970 Evo Plus). It's in a 2 drive mirror.

    I wonder if someone who's better with BTRFS can have a look, or point me in the right direction.

     

    Scrutiny is showing a ton of medium errors, but it's been doing this on both drives, from day 1. I updated firmware at the time, but it didn't seem to have any affect, and the drives have been running fine for well over a year at this point.

     

    image.thumb.png.6f8b5de35745a0cff039be5e599c9ad4.png

     

    Snippet:

    Apr 26 07:03:37 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812252712960 on dev /dev/nvme1n1p1, physical 469695516672, root 5, inode 1072832, offset 4437626880, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:37 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21284, flush 0, corrupt 0, gen 0
    Apr 26 07:03:37 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812249567232 on dev /dev/nvme1n1p1, physical 469692370944, root 5, inode 1072832, offset 4434481152, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:37 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21285, flush 0, corrupt 0, gen 0
    Apr 26 07:03:41 server kernel: nvme_log_error: 22 callbacks suppressed
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917376120, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: blk_print_req_error: 22 callbacks suppressed
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917376120 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: scrub_handle_errored_block: 14 callbacks suppressed
    Apr 26 07:03:41 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812252721152 on dev /dev/nvme1n1p1, physical 469695524864, root 5, inode 1072832, offset 4437635072, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917369976, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917369976 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: btrfs_dev_stat_print_on_error: 14 callbacks suppressed
    Apr 26 07:03:41 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21300, flush 0, corrupt 0, gen 0
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917379192, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917379192 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812249575424 on dev /dev/nvme1n1p1, physical 469692379136, root 5, inode 1072832, offset 4434489344, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:41 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21301, flush 0, corrupt 0, gen 0
    Apr 26 07:03:41 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812254294016 on dev /dev/nvme1n1p1, physical 469697097728, root 5, inode 1072832, offset 4439207936, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:41 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21302, flush 0, corrupt 0, gen 0
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917382264, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917382264 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917385336, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917385336 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: BTRFS warning (device nvme1n1p1): i/o error at logical 812255866880 on dev /dev/nvme1n1p1, physical 469698670592, root 5, inode 1072832, offset 4440780800, length 4096, links 1 (path: domains/eco/vdisk1.img)
    Apr 26 07:03:41 server kernel: BTRFS error (device nvme1n1p1): bdev /dev/nvme1n1p1 errs: wr 0, rd 21303, flush 0, corrupt 0, gen 0
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917388408, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917388408 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0
    Apr 26 07:03:41 server kernel: nvme1n1: I/O Cmd(0x2) @ LBA 917373048, 8 blocks, I/O Error (sct 0x2 / sc 0x81) MORE DNR 
    Apr 26 07:03:41 server kernel: critical medium error, dev nvme1n1, sector 917373048 op 0x0:(READ) flags 0x800 phys_seg 1 prio class 0

     

    Diags attached as well.

    server-diagnostics-20230506-1116.zip

  4. Well I'm looking at it because I'm maxing out CPU some of the time (4K/HDR/HEVC library) with enough users, so I'd be upgrading for perf reasons anyway, I'm just thinking to reduce power while I'm at it, given that having that CPU stuck at 100% for hours at a time doesn't feel great.

     

    I hear you on ROI. My server at the moment is costing around €35-40 per month to run, so if I could cut idle power by 50W and lots-of-plex-power by maybe 100W(?), the savings there would be enough for a couple of coffees a month, at least.

     

    Larger drives would be nice, but I've been continually expanding the server as the content grows, so I'll just be phasing out the 8TB drives with 18/20TB ones as needed.

  5. Thanks for the response. That's significantly lower than mine alright. I know the HBAs and NIC will be pulling a certain amount of power (and probably the backplane as well, come to think of it), but my idle draw still seems high, though I suppose that's what happens with a 4-year-old architecture at this point.

     

    I'd like to go intel for iGPU, and I like the idea of having game server stuff on performance cores, and unRAID stuff on efficiency cores. Only downside is LGA 1700 server boards are expensive, and there aren't many consumer boards that offer the layout or features that I'm after.

  6. Hi all,

     

    Electricicy getting more expensive these days, I'm taking a closer look at my server's current power usage, and eyeing up a more power efficient build.

     

    At the moment, my server idles at 140W. All drives spun down, nothing running. 170W with drives spinning.

    Hardware: 13 8-18TB drives, 2 SATA SSDs, 2 NVME drives, 2 9207-8i controllers, an X-540-T2 NIC, and a 3900X, running on an X470D4U

     

    The controllers use around 10W each, and the NIC can use around 15-20W. Let's assume another 5W each for the solid state storage, and that means my CPU is pulling around 80W at pretty much zero load, which seems high. I have turbo enabled, CPU governor set to power save.

     

    Can anyone with anything similar share numbers, in case I might have missed something here?

     

     

     

    Second, I'm thinking of switching to something like an i5 13500, for Plex to use the iGPU rather than CPU at the moment.

    What are people's thoughts on this? I assume power consumption is pretty low while transcoding but it's a difficult thing to find numbers on. Anyone have a better idea? 

  7. Not at the moment. 

     

    Your best bet is to create a new pool with the extra disks. 

    If you're on the stable version of unRAID, this will be a BTRFS pool, defaulting to RAID1, but this can be changed to RAID5/6 if desired.

    If you're using a beta version of 6.12, you can also choose a ZFS pool.

  8. Sounds like it'll be a more direct competitor to the QNAPs/Synologys out there. Interesting times.

     

    unRAID is amazing in its flexibility, but (I think) it's lacking in the polish department. 

    Things like downsizing or removing disks from the array, rebuilds (without having to remove parity first), a native backup solution, and general system monitoring, to name a few.

     

    It will be interesting to see where the focus is for this new product, and to see how they handle the more complicated stuff (if at all) that unRAID offers.

  9. Yeah sure, my bad!

     

    #!/bin/bash
    logger "SSD Pool > Array backup started"

    The inner-workings of very first line aren't super-important, just know that it should be at the start of every script you make for now.

    This first bit just sends a message to the unRAID syslog "SSD Pool...."

     

    # Backup things from SSD pool to array disks
    rsync -avh --del --stats \

    This next bit is just a comment, saying what it's doing.

    The next line is the start of the rsync command. In general, it looks like:

    rsync -someOptionsHere /path/to/source/directory /path/to/destination/directory.

     

    So with this one, the options are a, v, and h.

    a = archive. Basically what you want. Copy any new files from source, to destination, overwriting if the file is newer, but don't delete anything from either side. Also keep all the ownership, permissions, etc.

    v = verbose. Just log what it's actually doing while it's doing it, in case something screws up and I need to look at it later.

    h = human-readable. Just so it's in KB/MB/GB instead of bytes.

     

    Then the next bits are more options, but they don't have single-letter shorthand that can be combined together, so you have to specify them separately.

    --del = delete from source directory after copy to destination is finished. Because the whole point is we want stuff off the cache here.

    --stats = more info about what it did at the end of the operation, again for logs.

     

    The very last thing is the \ symbol. This basically says, continue the command, but on the next line. Just for my own readability.

     

    --exclude 'backup/' \
    --exclude 'downloads/' \
    --exclude 'movies/' \
    --exclude 'tv/' \
    --log-file=/mnt/user/backup/ssd-pool/backup-log.txt \
    /mnt/ssd/ /mnt/user0/backup/ssd-pool/

     

    These continue the options like before, this time to exclude certain directories from the copy, becuase I don't care about them much, so it's not the end of the world if they get lost. I mostly made this script to keep my documents and pictures a little more safe.

     

    The second-last line is an option to write all the logging we've been doing to a file. In this case "backup-log.txt" on the SSD.

    And the very last line specifies our source and destination. So my cache is called "ssd" so we have "/mnt/ssd" as the source, and I want it to go to a share called "backup" and a folder in there called "ssd-pool".

     

    The important bit here though is that we don't want it going to the user share (/mnt/user/backup) because this includes the cache and the spinning disks, so the files wouldn't actually move, or might even get corrupted, so we specify "user0" (this is a short-hand for "the share, but exclude cache").

     

    Hopefully all that made sense!

    • Like 1
    • Thanks 1
  10. 11 hours ago, trurl said:

    There is no such thing.

     

     

     

    8 hours ago, Hexenhammer said:

     

    Share that is only on the array and not on the pool?

     

    This is what I meant.

     

    8 hours ago, Hexenhammer said:

    Can you share the script please?

     

     

    Sure. You'll need to change the names of shares obviously, and I take no responsibility for your server catching fire, etc.

     

    #!/bin/bash
    logger "SSD Pool > Array backup started"

     

    # Backup things from SSD pool to array disks
    rsync -avh --del --stats \
    --exclude 'backup/' \
    --exclude 'downloads/' \
    --exclude 'movies/' \
    --exclude 'tv/' \
    --log-file=/mnt/user/backup/ssd-pool/backup-log.txt \
    /mnt/ssd/ /mnt/user0/backup/ssd-pool/


    logger "SSD Pool > Array backup finished"

     

  11. Has anyone much experience with the custom scripts before/after mover?

    I've got one I'm trying to run that stops/starts my download client before/after.

    Rather than two separate scripts, I'm passing the same one with a viable, so In the two fields, I have:

     

    image.png.b06d3829bea4353b13df714a4abe265b.png 

     

    Looking at the log file, the variable is empty, so it's not being passed through.

    I can make a second one, but maybe a feature request would be to alter that field to allow more than just a path?

  12. To be fair, this has been requested before, and I know I'm not the only one that's implemented this as a script.

    The standard answer is it's not an option because of how mover runs, but I don't think it's too much of an issue to add this functionality, with a big red warning to use array-only shares (or even just don't allow the destination share to be selected at all if it's set to use cache)

  13. I think this has been asked before, but it was a while ago, so forgive me if you end up repeating yourself:

     

    Is there an option to have certain shares ignore tuning rules, but still get moved on schedule?

    There's an option to have files on a file list excluded, but this seems to exclude them from the mover entirely.

     

    My use-case is that I want my backups going to my cache for speed, but I don't want them sitting there for three days taking up space that could be used for actually frequently accessed files.

  14. +1

     

    I'm doing this manually at the moment, with a baseline VM that I copy vdisks from to a new VM's folder, than manually edit settings.

     

    Ideally, clicking "Clone" would bring you to a VM creation page, with all details filled out to match the previous VM.

    Starting at this point would give some sort of error, telling you to change at least the name, I guess.

    • Like 1
×
×
  • Create New...