Jump to content

Jclendineng

Members
  • Posts

    201
  • Joined

  • Last visited

Posts posted by Jclendineng

  1. 53 minutes ago, doron said:

    @Jclendineng, if I was unclear, that was a genuine question. Since you suspected a bit flip, that would be a viable hypothesis only if you have exactly one data drive in the array. Each drive has its own LUKS header; the chance of a bit flipping in all of them at once is practically zero.

    I have 1 data drive, but I also have encrypted cache drive, with its own header, the keyfile doesn't work for either header anymore, but the header itself doesnt look corrupted meaning something must be wrong with the keyslot...in any case this is on me for not backing up the header.

     

    Edit. 1 data, 1 parity and 1 cache.

  2. 9 hours ago, doron said:

    Need some context - have you been using the script that's the subject of this thread?

    Yes, though I think I found the issue, or at least the most logical one. I don’t have ecc ram in my current server and a cosmic ray or random bit flip happened hosing my key slot. My key slot looks fine with the checker and the luks header looks fine so the only thing it could be is 1 bit maybe got changed causing my keyfile to fail. Unfortunate and I’ll have to format all my drives now but it happens and apparently in my research on this, happens a lot. If even 1 bit fails on the key slot the header is hosed. There are a couple threads on consistently backing up the header with mixed results, unfortunately I didn’t have a good header backup so all my data is gone. Sucks, but it’s price you pay for encryption, especially when it’s not fully supported by unraid. Future me will keep up to date header backups. 
     

    Edit. I think it would be handy to have the script give the option to use more than 1 key slot when updating the key, instead of always deleting the old key by default, so if someone wanted to use a pass phrase and a key file for example they could. Unrelated to this but it would be a good addition to an already nice script (until unraid makes it officially a part of the gui as limetech has stated they will do)

  3. On 2/18/2023 at 8:44 AM, Gico said:

    Hi. I scheduled parity check for the first time, with "Use increments for scheduled Parity Check" enabled,

    and "Write corrections to parity disks" set to to "Yes".

    This morning I got a notification that "[Juno} paused Scheduled Non-Correcting Parity Check (3.2% completed)".

    Why does this notification says "Non-Correcting"?

    I would turn incremental off, I enabled it, and when my monthly parity check was running it kept stopping (as expected) but then it completely freaked out, saying it couldn't write to the disk etc (new disk, cleared and all that), and unfortunately killed all the data forcing me to rebuild from parity. I do have parity so I didn't lose anything but be warned that something may not be completely stable here. Have a backup. 

  4. What sort of ZFS data recovery/protection features can we expect from the GUI, I know ZFS has self healing but if a drive fails and needs replaced in a raidz configuration, will the GUI support basic hdd replacement and monitoring? What about zfs errors. Will those be exposed in syslog?

  5. I have a new server set up to test the zfs stuff, I 100% wont be doing testing on my main server 😂 then if I can get raidz working as expected then I’ll encrypt to test that stability and let it bake a bit after final release and copy my data over. I felt like ZFS official support deserved a server upgrade and it’s a good time to do it since I’ll need to reformat anyways for zfs. 

  6. 42 minutes ago, GRRRRRRR said:

    So you need additional protection from unRAID Parity ontop of what ZFS already does for protection?

    Sounds kinky and experimental. Better respect the filesystem authors intentions and don't mix the two teams. I think the mix is already called btrfs. Depending on what you need that protection for, I am sure if you explain it, we can come up with the best safest and so on solution, without having to "virtualize" filesystems ontop of eachoter. Been there done that, it could work, but... not in production.

     

    The best protections are designed and pre-configured in a way that takes into account all combos between all fail scenarios across the protection layers. Failure mode and effects analysis is the process of reviewing as many components, assemblies, and subsystems as possible to identify potential failure modes in a system and their causes and effects. The issue is that how each layer works it's either guaranteed 100% and documented by the authors and you can understand and plan for it OR you will have to read the source code to fully understand and anticipate the logic (simulate the software in your head X the other layer X all failure modes permutations) ?

     

    Just tell us what you mean by that, why do you need that additional Parity 

    I was just being dumb. RAIDz1-3 would do what I want, I had a brain fart. So 3 drives in raidz1 would be best storage to parity efficiency 

  7. So I’m thinking about doing a flash based zfs array with this. How does parity with with ssds? TRIM and all that. Will a flash based zfs array in raidz work with parity? So 3 ssds plus a large platter drive for parity. The ssds will make up the zfs array and parity would be platter based. Would that be possible with zfs?

  8. Following, in the exact same boat. I have a r630 I want to move to but don’t want to fill with hdds as ssds are more power efficient and a lot faster plus less moving parts. Unsure how unraid handles ssd’s in an array and if I can even use ssd’s with parity. Need to figure it out, or I’ll just use bare metal podman or rancher or something to manage docker. I couldn’t find anything conclusive on the forums about ssd’s and parity. 

  9. More clarifications:  in Unraid OS only user-defined pools can be configured as multi-device ZFS pools.  You can select ZFS as the file system type for an unRAID array disk, but will always be just a single device.  The best way to think of this, anywhere you can select btrfs you can also select zfs, including 'zfs - encrypted' which is not using zfs built-in encryption but simply LUKS device encryption.
     
    Also note that ZFS hard drive pools will require all devices in a pool to be 'spun up' during use.  IMO where ZFS will shine is in large flash-based pools (SSD, NVMe, etc).

    Interesting I was going to ask about encrypted zfs since I use encrypted xfs currently. So I’m curious why not zfs native? Isn’t it easier to manage then LUKs?


    Sent from my iPhone using Tapatalk
  10. 20 hours ago, jackfalveyiv said:

    I replaced my RAM last week after some errors were occurring.  Today after backups ran I woke up to errors due to memory.  Do I need to add more?  Is there an errant backup process mucking things up?  Thanks in advance.

    trescommas-diagnostics-20230102-1021.zip 867.33 kB · 0 downloads

    I would update unraid first and post diagnostics when it happens again, looks like you are on 6.11.1, 6.11.5 is the latest bug fix release and contains a lot of improvements. there were numerous issues with 6.11 that have been subsequently fixed in point updates, that's where I would start. I also see some virtio errors that probably aren't related but usually occur when there is a GPU or other device that's not being passed correctly to a VM. The app data backup/restore issue is also separate and you would ask in that thread, but if you have error checking on the backups that could cause that error since (depending on what you are backing up) it will always fail that (as an example if you backup USB or system, since checksums will fail as the files will have changed between start and end of backup)

  11. On 12/21/2022 at 9:05 AM, b3rs3rk said:

     

    If you're happy with staying version locked forever, cool.  But if you want help fixing the issue I will need a lot more detail starting with everything I outline in the troubleshooting section of the first post of this thread.

    I am, its just utilization, but I have updated again and included the troubleshooting steps listed in the OP, thank you!

     

    Edit: I updated and it works...Im not 100% sure why. Possibly the APU needed reset, but I had thought that last time and rebooted which should reset it, which didn't fix.  Ill reply with drag info if it pops up again...Thank you!

  12. I am getting "Multiple NICs on the same IPv4 network"

     

    eth3 and eth7 both have IP addresses on the 10.0.2.0 network. This is rarely a valid configuration.

     

    I don't have an eth7 anymore, I did throw a NIC in to test something but I didn't assign any IP address and its been removed.  I already deleted and regenerated the docker config file and network rules file to try and solve but the warning won't go away. Attached are diagnostics. hank you for all your work!T

    tower-diagnostics-20221226-0851.zip

  13. How use a secret as enviroment variable in a Docker/Template?

    No way to do it currently so I don’t use it, I don’t want to manually unseal every time I patch unraid unfortunately


    Sent from my iPhone using Tapatalk
  14. The only good DIY alternative would be to proceed ZFS replications and using Mikrotik router (cheaper and customizable) to proceed "heartbeat" from your VM with ping or from your host each minute. In case of failure, of VM or container answer, the Mikrotik can proceed an SSH session to boot the sibling VM/container on the backup server, but trafic also need to be manage on the network side.
     
    So another mess to script also according to your need.

    That’s true, I could do that, I was more thinking about ha within unraid so running 2 instances of the sql docker so I can patch one and not take all the apps relying on it down.


    Sent from my iPhone using Tapatalk
  15. What is HA?... I've only heard of it as "High Availability" which is kind of redudndant when talking about servers.
     
    What are you trying to do?
     
    6.
     
     

    High availability and no, it’s not.


    Sent from my iPhone using Tapatalk
  16. Seems that no one uses HA in unraid...I will reply to this if/when I find a good solution, I use LinuxServer MariaDB and upgrading the DB requires user interaction, and it really needs to be HA as its a DB that many apps rely on. I know Maria Enterprise has HA features. 

  17. Version 2022.11.30a I get N/A for everything as well, just adding that here.  I have had this for a while and thought it was just something I did, good to see its not just me.

     

    Edit: Rolled back per the person a few posts back and it works, thanks!!

     

    https://raw.githubusercontent.com/b3rs3rk/gpustat-unraid/6cf1b1e96bc8cd5c1cf7ac8fefea1271d8891e26/gpustat.plg

×
×
  • Create New...