HNGamingUK

Members
  • Posts

    134
  • Joined

  • Last visited

Posts posted by HNGamingUK

  1. On 12/14/2021 at 5:52 PM, bonienl said:

    The plugin Dynamix Stop Shell can help mitigate a hanging system. it will terminate all open shells when the array is stopped.

    Keep in mind that forcibly terminating a process may lead to data loss.

     

    The plugin would work yes however it automatically stops the sessions without any user input or knowledge. 

     

    My feature request is for a prompt and list so the user has full knowledge of the process(es) stopping shutdown of the array and can action it all within the unraid UI. 

  2. It was mentioned today on the unoffical Discord that when you shutdown the array it has the potential to hang and does not inform you of any open processes and expects you to find them and close them before it will finish the shutdown. 

     

    This request is therefore to add a prompt in the GUI (maybe also CLI) that when a user issues a command that will shutdown the array to display a message to the user. This message would be something to the affect of "The following processes are stopping shutdown of the array" at which point it lists the processes and PIDs. (Likely will need to have this auto refresh as other processes could start to use the system)

     

    In addition to the simple message, it could include a prompt asking the user if they would like to kill the processes and force the array shutdown. 

     

    I believe adding such a feature will be a very good addition and improve the overall user experience. 

  3. 1 minute ago, wgstarks said:

    The first two aren't important. If you scroll up a little you'll see that there has been a bit of discussion about this today. I believe most people solved the last one by switching endpoints.

    Yeh I saw the only thing is that due to this being a new setup there is no wireguard .conf file in the wireguard directory for me to be able to change the endpoint....

  4. Hello I am trying to start a container from fresh using wireguard and PIA but I am getting the following in the docker logs?

     

    2021-04-15 00:15:40,475 DEBG 'start-script' stderr output:
    parse error: Invalid numeric literal at line 4, column 0
    
    2021-04-15 00:15:40,600 DEBG 'start-script' stderr output:
    parse error: Invalid numeric literal at line 1, column 7
    
    2021-04-15 00:15:40,600 DEBG 'start-script' stdout output:
    [warn] Unable to successfully download PIA json to generate token from URL 'https://143.244.41.129/authv3/generateToken'
    [info] Retrying in 10 secs...

     

    I assume this just means it's an issue with PIA and I just have to wait? But was also concerned about the previous two parse errors...

  5. Hello Limetech,

     

    Talking in the unoffical discord we noticed when you add a new container and click apply it runs a docker run however when you edit a stopped container it does a docker create

     

    Now this method is okay however some people would like the option to not have a container to start on creation.

     

    Essentially the feature request is to add a tickbox called something like "Start container on creation/edit" on a container creation and edit page. By default this would be ticked so that it does not affect the current usage of the containers and does not cause any confusion.

     

    Potentially in addition an option within the Docker settings page could be added to change how the tickbox works by default.

     

    Hopefully this makes sense.

    • Like 1
  6. This has been requested atleast once before by @aptalca back in 2018 however this still doesn't seem to have been implemented?

     

    Basically instead of one tarball of all the appdata directories, please could we have the option to have it seperated per directory? This will make a restore much easier for people who only want to restore one application. Currently they would have to manually untarball the file and move the directory.

     

    So the feature request is:

     

    Backup:

    In the backup config page have a option called "Backup Directories Seperately"

    If this option is selected it will backup each directory in the provided appdata directory into a time and dated folder within the provided backup directory.

    In addition to this if the "Compression" option is ticked then compress each directory.

     

    Restore:

    When someone chooses the restore option, along with the date/time selection provide a directory selection to restore (using checkboxes to allow for multiple selections and an "all" option to restore everything)

     

    Hopefully this is enough detail for the feature request to be actioned... @Squid

    • Like 4
  7. 3 hours ago, Phaelo said:

    Restarted array and I was able to get the SMART data on the drive now, attached.

    donnager-smart-20201205-2231.zip 3.84 kB · 1 download

    Reading the SMART data suggests the disk is fine, I would first check the cables to the disk (power and SATA)

     

    If everything looks fine you should be able to:

    1. Stop Array

    2. Un-assign drive

    3. Start Array

    4. Stop Array

    5. Assign Drive

    6. Start Array

     

    Once this has been complete the drive should be added back to the array and any data changes that have happened to the emulated drive in the meantime will be written back to it.

  8. 9 minutes ago, PeterNet said:

    Well, then it no longer works as it used to do in v5.

    Turning on the cache, it was transparent with the Disk Array share.

    I just saved files to:
    192.168.x.x\sharename\folder1\folder2\etc...

     

    But instead of being placed directly on the main array, the files were placed on the Cache disk (with a copy of the folder structure).

    Then doing the Mover, it would transfer everything to the actual Disk Array.

     

    So that is no longer possible then?

    I would check the "Cache" setting of the share in question and confirm that it hasn't been set to "No"

  9. Hello,

     

    So as the title suggests I am unable to access my dockers that are in default bridge mode from a Windows VM (on unraid) from a VLAN (br0.20)

     

    Here is what my unraid network config looks like:

    image.thumb.png.ac659c092941de492a3fc8e76b7f2505.png

     

    Here is a screenshot of my dockers :

    image.thumb.png.f55ecc88f0b29da840d1dca696b997e0.png

     

    While on my Windows VM (which has it's network as br0.20) I am unable to access the docker containers on the standard bridge mode (eg Grafana) however I am able to access the dockers using custom networks (eg UNMS).

     

    My network setup as follows:

    br0 (Standard bridge) is on a 10.0.1.0/24 subnet

    br0.20 (VM assigned) is on a 10.0.2.0/24 subnet

    br0.5 (2 of the dockers) is on a 10.0.0.0/24 subnet

     

    I am able to access any of the dockers webUIs from a mobile device on the same subnet as my Windows VM. I believe this has something to do with the network isolation between dockers and VMs am I right?

     

    How would I go around allowing access from the Windows VM to the dockers on the default bridge?

     

  10. 11 minutes ago, Squid said:

    It calls the update script to check for and install updates at the time the backup runs

    Okay brilliant so setting the following in Auto Update:

    image.thumb.png.f7cb151e773f536dcb8d30f49f030d5a.png

     

    and then the following in backup:

    image.thumb.png.d5b05f60ea0d93e65f900e4307ed74b5.png

     

    Will backup every Sunday at 3:05 and once complete trigger an update.

  11. Just now, testdasi said:

    Indeed the limit is for the kernel to know when to kill a docker.

    I don't think the docker itself is even aware of the limit (perhaps it can be aware but I don't think that piece of code implemented in all the dockers that I can tell).

     

    There's also a side effect for setting docker memory limit is that it is less likely to kill your VM if you actually run out of memory. So don't just go undo all your docker limit.

    Yeh I didn't want my VM to die randomly as I use it for gaming.

     

    I will have a look at working out some better docker memory limits. Thanks for helping me understand how it works.

  12. 7 minutes ago, testdasi said:

    The 70% utilisation you see right now is after the offending process has been killed so it's likely a red herring.

     

    OOM error happens if (a) you actually run out of memory or (b) a docker runs over its memory limit or (c) unusual problem

     

    Did you set the dockers memory limits before or after mysql docker was killed?

    If before then you need to increase the limit for mysql as it was killed because of reason (b) above.

     

    And OOM is not an urgent issue to fix as long as it doesn't kill your running docker / VM.

     

     

    Side note: Reason (c) is really unusual and not something you need to worry about. For example, my TR4 server with F12 BIOS will have OOM error if I allocate more than 50% RAM all at once (e.g. start a VM) while having Global C State disabled (enable C state and no more error).

     

    Yes I set the MySQL and other docker limits before MySQL was getting killed.

     

    I assumed that if the container got close to the limit it would just not use more, but reading your messages suggests that if it goes over the limit I set it will just kill the docker?

  13. Hello,

     

    For reference my server has 32GB of RAM.

     

    I wonder if anyone else is able to help me, I am unsure on what is causing this and what I can do to stop it.

     

    So far for the past couple of nights I have had fix common problems send an alert about out of memory errors and when I login I see that the MySQL docker has stopped. (Only this one stops so I am guessing OOM killer decides to kill that one for whatever reason)

    image.thumb.png.9865f2663d44d7124fe48e99fed3e6c4.png

     

    However my memory on the dashboard is hovering around 70% max so I am unsure on what is causing it.

    image.png.62d7cc51f74a224f936c02db137b744e.png

     

    I have set each docker to have max memory so that they can't randomly use more memory, however it still keeps happening.

    image.thumb.png.64ace3c3ff795e05e18732e8825d5e1c.png

     

    The only other item that I run contstantly is a Windows VM (16GB assigned to it) and as such should leave 16GB for the dockers and system.

     

    I have also attached my diagnostics, if anyone can help me find out what is causing this it would be wonderful. I don't have the capcity to just upgrade my RAM currently and so need to fix this issue. diagnostics.zip

     

  14. 7 hours ago, trurl said:

    Okay great that information helped.

     

    As I had the secondary cache working and all the data was there I just copied it from the single drive cache pool to the array. I then stopped the array, assigned the primary cache back and started the array. I then ticked the format box and started the format of the primary cache.

     

    As expected this formatted the cache pool, so all I had to do is restore the files to the cache and everything was back to normal.

     

    Just so I know for future, reading the errors shown can someone explain what happened?

  15. 24 minutes ago, jonathanm said:

    It would appear the qm command is specific to proxmox, not a generic QEMU-KVM command.

    Yeh it seems that way, unsure if it is possible to do with the current setup unraid/lime tech use. But it would be a nice feature to allow live disk expansion as most newer operating systems would support it.

  16. Hello,

     

    Wondering if anyone can help me with an issue I am having with my Cache pool. Basically after a reboot today, when I started up the array I saw the following:

    84426140_CacheDrives.thumb.png.ba0f09b5d6be410d1d997a95bdde5c94.png

     

    Along side this I see the below asking me if I want to format the primary cache drive. Now obviously I do not want to as I don't want to loose data.

    Format.thumb.png.09f85fed9ad155a722913226cc26a5d2.png

     

    I then ran a btrfs check --readonly /dev/sdb1 which came with the following output:

    Opening filesystem to check...
    Checking filesystem on /dev/sdb1
    UUID: ab81d341-8531-4c08-8fa1-645911b301fd
    cache and super generation don't match, space cache will be invalidated
    found 310226411520 bytes used, error(s) found
    total csum bytes: 0
    total tree bytes: 90783744
    total fs tree bytes: 29900800
    total extent tree bytes: 60588032
    btree space waste bytes: 16964433
    file data blocks allocated: 156191051776
     referenced 154714845184

    I then did a btrfs check --repair /dev/sdb1 which showed the following:

    Starting repair.
    Opening filesystem to check...
    Checking filesystem on /dev/sdb1
    UUID: ab81d341-8531-4c08-8fa1-645911b301fd
    [1/7] checking root items
    Fixed 0 roots.
    [2/7] checking extents
    incorrect offsets 12845 12358
    incorrect offsets 12845 12358
    incorrect offsets 12845 12358
    incorrect offsets 12845 12358
    Shifting item nr 94 by 487 bytes in block 1254219743232
    Shifting item nr 95 by 487 bytes in block 1254219743232
    Shifting item nr 96 by 487 bytes in block 1254219743232
    Shifting item nr 97 by 487 bytes in block 1254219743232
    Shifting item nr 98 by 487 bytes in block 1254219743232
    Shifting item nr 99 by 487 bytes in block 1254219743232
    items overlap, can't fix
    check/main.c:4336: fix_item_offset: BUG_ON `ret` triggered, value -5
    btrfs[0x42f27d]
    btrfs[0x43842d]
    btrfs[0x438960]
    btrfs[0x43950c]
    btrfs[0x43d495]
    btrfs(main+0x90)[0x40ecc0]
    /lib64/libc.so.6(__libc_start_main+0xeb)[0x1473a7ed6e5b]
    btrfs(_start+0x2a)[0x40ef4a]
    Aborted

    So at this point I am stumped on what to do to resolve the issue.

     

    I have been able to set the primary cache as no device and then start the array at which point it seems the secondary cache drive is working fine and has all the data I need. However preferably I do not want to run at a reduced redundancy for an extended period of time.

     

    If I am infact being stupid and do need to format the primary cache drive to then make it re-sync the raid I can do that. Any help infact would be very much apprriciated.

     

    Appologies this post is VERY long, I just wanted to make sure that I covered all the options I could find before asking for help.

  17. On 4/5/2020 at 2:02 PM, jonathanm said:

    Can you link to official documentation that walks through the live expansion? All I can find are work arounds and hacks that come with data loss warnings. I guess my google-fu is failing me.

    For Proxmox here is the documentation: https://pve.proxmox.com/wiki/Resize_disks

     

    Although looking further it may seem that Proxmox uses a different vdisk handling method not sure myself. I just know it is possible with Proxmox both in the GUI and via CLI

  18. 10 hours ago, jonathanm said:

    Where are you seeing that it's even possible to expand a KVM guest vdisk without stopping it? The googling I've done suggests it's required to stop the guest.

    https://computingforgeeks.com/how-to-extend-increase-kvm-virtual-machine-disk-size/

     

    When the guest is stopped, it's easy to expand the disk from the GUI, on the VM page click on the NAME of the VM, and click on the disk capacity. Enter the new larger number with G modifier and hit enter.

     

    NEVER EVER EVER set a smaller size. I know that's not what you asked, I'm just posting as a reference if someone searches and finds this answer thinking they can shrink a vdisk just as easily. You can, but you will break the vdisk permanently, and probably not be able to recover your data.

    Proxmox would be my example of live disk increase, it uses QEMU KVM.

  19. Hello All,

     

    I was wondering if someone will be able to help me. I am running a Windows VM and preferably don't want to have to turn it off to increase the disk.

     

    From what I can see it can't be done via the GUI? Is this a new feature that needs to be requested?

     

    In either case what would be the best command to run to increase the disk?

     

    Many Thanks

  20. Hello,

     

    Is anyone having issues with PIA (specifically the france enpdpoint) I am getting the following:

    2020-03-31 07:16:56,249 DEBG 'start-script' stdout output:
    [info] PIA endpoint 'france.privateinternetaccess.com' is in the list of endpoints that support port forwarding
    
    2020-03-31 07:16:56,249 DEBG 'start-script' stdout output:
    [info] List of PIA endpoints that support port forwarding:-
    [info] ca-toronto.privateinternetaccess.com
    [info] ca-montreal.privateinternetaccess.com
    [info] ca-vancouver.privateinternetaccess.com
    [info] de-berlin.privateinternetaccess.com
    [info] de-frankfurt.privateinternetaccess.com
    [info] sweden.privateinternetaccess.com
    [info] swiss.privateinternetaccess.com
    [info] france.privateinternetaccess.com
    [info] czech.privateinternetaccess.com
    [info] spain.privateinternetaccess.com
    [info] ro.privateinternetaccess.com
    [info] israel.privateinternetaccess.com
    [info] Attempting to get dynamically assigned port...
    
    2020-03-31 07:16:56,254 DEBG 'start-script' stdout output:
    [info] Attempting to curl http://209.222.18.222:2000/?client_id=
    
    2020-03-31 07:16:56,282 DEBG 'start-script' stdout output:
    [warn] Response code 000 from curl != 2xx
    [warn] Exit code 7 from curl != 0
    [info] 12 retries left
    [info] Retrying in 10 secs...

    Settings:

    518426115_DelugeVPNSettings.thumb.png.ee07b9a7fa499f993a7cdef7ba214039.png

  21. 1 hour ago, Turrican said:

    Thanks, ill try a new provider. Will go either Mullvad or Surfshark, unless anyone has anything bad to say with them for this setup.

    May I ask why you are going for the smaller brand VPN providers as apposed to say PIA, AirVPN or NordVPN?