Nexius2

Members
  • Posts

    152
  • Joined

  • Last visited

Posts posted by Nexius2

  1. Hello,

    I have a parity disk on my server and I want to add a second one.

    what is the process knowing that I can stop / restart array but it must not be offline for hours. meaning during rebuild, the array MUST be online with all VMs and dockers running.

    I also will have to change the actual parity disk with a bigger one

    thanks

  2. Great question same conf on my side and about to do the same thing.

    what I have tested for know:

    - stop array

    - remove parity dsk 1

    - start array

    - stop array

    - add parity disk 1

    - start array

     

    My array is encrypted so it might be the cause (or the parity disk 2 but I doubt), but while rebuilt, the shares are not mounted so array not usable.

     

    if you find a way to keep the array usable (and VM and Docker) while rebuild / add parity, I'll take it

  3. so, I added disk20 a couple days ago. no hot plug so I stoppd the server to add it. array was fine nothing to notice after array growth.

    and 1 or 2 days after the parity check started (it stops at 18h and starts back at midnight).

    during the check it seems to have found errors on disk 19. since then, parity check has stopped, but I can see data move between all the HDD as if it was still going on.

    I just notice sdac has appeared (thanks to you) don't know if it's good, it will be changed anyways.

    maybe some power plug has moved while adding disk20....

     

    but nothing realy explains why parity is stuck at 19.8%

     

     

  4. Hello,

    I updated one of my servers last night from 6.11.2. didn't have issue before, no changes done recently and this morning I have "Invalid folder Backups contained within /mnt"  ( i posted in "[Plugin] CA Fix Common Problems" but I realize i might have some reel issue) and "Rootfs file is getting full (currently 83 % used)" error messages.

    all i can find about this is from RAM being full, but I have 48Gib and only using 28%.  i'm guessing I have a reel issue because when I try to move a file (10gb) from a share to another (both on the same cache drive) it takes a couple secondes, just like it would between 2 drives. yesterday, it was instantly done.

    I have a script that mounts to another server and rsync data every night. the receiving server whas off tonight, so maybe it could have tried to copie data in /mnt/backups but I can't find anything there.

    any idea?

    nostromo-diagnostics-20221110-0716.zip

  5. Hello,

    I'm starting to need more space and looking at new HDDs.

    I had a bad experience with seagate ironwolf (8To) on my HBA card that failed randomly. No bios update on conf ever manage to solve my issue so I bought a simple sata controleur for them and continued with WD on my HBA.

    now, I'm looking at seagate exos X18 18To but I don't want to buy them if I can't connect them to an HBA card.

    has anybody tryied this with issue? (I meen for more then a month).

    thanks

  6. Hello,

    comming again for some help 🙂

    since I had no logs starting my container, I tryied with command lines. this is what I got

    docker run plex_user_manager
    Unable to find image 'plex_user_manager:latest' locally
    docker: Error response from daemon: pull access denied for plex_user_manager, repository does not exist or may require 'docker login': denied: requested access to the resource is denied.

    My dockerfile seems ok, I have it synced to dockerhub succefully, so it should be ok.

    here is my template

    https://github.com/Nexius2/Unraid_templates/blob/main/plex_user_manager.xml

     

    any idea what I do wrong?

    thanks

  7. today, I had a script use a share mount. in fact, it's the "backup/restore appdata" plugin that used a unassigned device mount to backup. and it fails.

    there is something not working well with unraid mounts. when I search on the forum I see lot's of "kernel: traps: lsof[****] general protection fault ******* in libc-2.36.so" and other similar errors.

     

    I thought my servers where just failling because of too long response because of high cpu usage, but realy I'm begining to doubt.

     

    what is, the best practice to make shares on unraid?

     

  8. 6 hours ago, dlandon said:

    Looking at your diagnostics I see an issue with one server:

    Oct 10 08:56:15 Halcyon unassigned.devices: Warning: shell_exec(/bin/df '/mnt/remotes/AURORA_tdownloaded' --output=size,used,avail | /bin/grep -v '1K-blocks' 2>/dev/null) took longer than 5s!

    This is generally indicative of network or remote server connection issues.

     

    That server is having a tough time with a CIFS mount:

    Oct  9 06:46:29 Halcyon unassigned.devices: Mount SMB share '//AURORA/tdownloaded' using SMB 3.1.1 protocol.
    Oct  9 06:46:29 Halcyon unassigned.devices: Mount SMB command: /sbin/mount -t cifs -o rw,noserverino,nounix,iocharset=utf8,file_mode=0777,dir_mode=0777,uid=99,gid=100,vers=3.1.1,credentials='/tmp/unassigned.devices/credentials_tdownloaded' '//AURORA/tdownloaded' '/mnt/remotes/AURORA_tdownloaded'
    Oct  9 06:46:29 Halcyon kernel: CIFS: Attempting to mount \\AURORA\tdownloaded
    Oct  9 06:46:29 Halcyon kernel: CIFS: VFS: Error connecting to socket. Aborting operation.
    Oct  9 06:46:29 Halcyon kernel: CIFS: VFS: cifs_mount failed w/return code = -111
    Oct  9 06:46:29 Halcyon unassigned.devices: SMB 3.1.1 mount failed: 'mount error(111): could not connect to 192.168.1.70Unable to find suitable address. '.

     

    What is that server?

     

    I would not mount that server share with UD and see if it stops your SMB issues.

    Aurora is a 3rd server and this one does not unmount from Halcyon (or at least mounts back after). I would guess errors are do to high CPU usage that stalls the server. Aurora is pretty much alway between 90 and 100% CPU 🙂

     

    My issue is from Halcyon and Nostromo or from Aurora and Nostromo because my shares are on Nostromo (and he is rarely over 40% CPU)

     

     

    But maybe I'm wrong and I have some sort of network issue between all my servers 😕