• Posts

  • Joined

  • Last visited

Posts posted by Mizerka

  1. Quote

    MikroTik RB951G-2HnD

    that looks retro


    anyways, tried applying those changes?


    since you can ping it and it echo's back then that's good enough, from here on, it'll be a layer 4 issue onwards, i'd still wager it's a microsoft service issue, got any other machines on network that can access this share btw?


    also looking at logs briefly, you have nameserver set to, typo or is that some strange public resolver you use? doesn't actually respond on 53 by the looks.



  2. 4 minutes ago, Frank1940 said:

    I went through your syslog and I am lifting out a section of it:


    Dec 18 15:39:15 FTP root: Starting Samba:  /usr/sbin/smbd -D
    Dec 18 15:39:15 FTP root:                  /usr/sbin/nmbd -D
    Dec 18 15:39:15 FTP root:                  /usr/sbin/wsdd
    Dec 18 15:39:15 FTP root:                  /usr/sbin/winbindd -D
    Dec 18 15:39:15 FTP emhttpd: shcmd (20608): smbcontrol smbd close-share 'FTP'

    The first four lines are the normal starting of the SMB process.  They occur whenever SMB restarts.  The problem appears to be in this fifth line.

    Dec 18 15:39:15 FTP emhttpd: shcmd (20608): smbcontrol smbd close-share 'FTP'

    Googling      smbcontrol smbd close-share   is a command that apparently orders smbd to close the client connections to the named share (FTP in this case).   Problem is that I don't have the foggiest idea what is generating this message. 


    I'm pretty sure that's the default behaviour when making changes to a share that's already deployed. it kills off connections and relies on client to reestablish it.


    looking at my logs;


    it's specifically ran when changing smb security settings;


    below changing public to secure



    Dec 18 23:00:18 NekoUnRaid root: Starting Samba: /usr/sbin/nmbd -D
    Dec 18 23:00:18 NekoUnRaid root: /usr/sbin/smbd -D
    Dec 18 23:00:18 NekoUnRaid root: /usr/sbin/winbindd -D
    Dec 18 23:00:18 NekoUnRaid emhttpd: shcmd (509): smbcontrol smbd close-share 'ssd'


  3. hmm those look fine, and you don't need dhcpv6 for ipv6 devices to work, apipa v4 169.254 is a pain to use, but ipv6 fixes that and can manage itself on networks without dns quite painlessly, win10 prefers it by default.


    anyway, doubt it'll be it, but you can try disabling ipv6 on your client interface to force it on v4, but network discovery worked fine over v6 by the looks.





    default eth int, and untick ipv6.


    also you don't have any firewalls/av in place that might prevent it? try disabling win firewall (temporarily if you actually use it)




    also make sure you have enabled discovery and filesharing on network/s (just add to all for now), found in network and sharing center.

  4. Hey,


    bit of a strange one, been on and off upgrading and working on unraid (in middle of encrypting entire array), over last week it crashed twice on me, I say crashed, but it's actually sitting at 80-100% cpu usage across all cores which it never does (24 cores, only running few dockers) and even more interestingly it's running 100% ram.


    Yesterday, I tried a few things like killing docker, force restarting, shutting downa array etc, but nothing worked, webui was somewhat usable and could console on to it as well.


    output of top:


    top - 19:23:29 up 22:10,  1 user,  load average: 154.26, 151.66, 147.51
    Tasks: 1062 total,   3 running, 1059 sleeping,   0 stopped,   0 zombie
    %Cpu(s): 19.1 us,  4.9 sy,  0.1 ni,  7.8 id, 67.4 wa,  0.0 hi,  0.8 si,  0.0 st
    MiB Mem :  96714.1 total,    520.3 free,  94856.5 used,   1337.3 buff/cache
    MiB Swap:      0.0 total,      0.0 free,      0.0 used.    223.3 avail Mem 

      PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+ COMMAND                                                                                                                       
     1809 472       20   0  153908  20228      0 D 501.0   0.0 301:48.74 grafana-server                                                                                                                
     1835 root      39  19       0      0      0 S  16.5   0.0  17:38.78 kipmi0                                                                                                                        
    25784 root       0 -20       0      0      0 D  10.2   0.0  42:59.54 loop2                                                                                                                         
      914 root      20   0       0      0      0 S   6.6   0.0  17:07.15 kswapd0                                                                                                                       
    11948 root      20   0    3424   2308   1680 R   5.0   0.0   0:00.15 lsof                                                                                                                          
     5922 root      20   0 5413656  61964      0 S   4.6   0.1 253:53.78 influxd                                                                                                                       
     6737 root      20   0  175536  18744      0 S   3.0   0.0 283:15.09 telegraf                                                                                                                      
      915 root      20   0       0      0      0 S   2.0   0.0   8:00.16 kswapd1                                                                                                                       
    12001 root      20   0       0      0      0 I   2.0   0.0   1:44.95 kworker/u50:0-btrfs-endio                                                                                                     
    17414 root      20   0       0      0      0 I   1.7   0.0   0:10.87 kworker/u49:0-btrfs-endio                                                                                                     
    18371 root      20   0       0      0      0 I   1.7   0.0   0:36.69 kworker/u49:3-btrfs-endio                                                                                                     
    25464 root      20   0       0      0      0 I   1.7   0.0   0:06.19 kworker/u49:5-btrfs-endio                                                                                                     
     9124 nobody    20   0  150628  11848   6372 S   1.3   0.0   4:15.60 nginx                                                                                                                         
     9652 root      20   0       0      0      0 I   1.3   0.0   0:33.14 kworker/u50:7-btrfs-endio                                                                                                     
    16159 root      20   0       0      0      0 I   1.3   0.0   0:27.58 kworker/u50:6-btrfs-endio                                                                                                     
    23079 root      20   0       0      0      0 R   1.3   0.0   0:39.64 kworker/u49:4+btrfs-endio                                                                                                     
    10860 root      20   0       0      0      0 I   1.0   0.0   0:03.74 kworker/u49:11-btrfs-endio                                                                                                    
    10955 root      20   0       0      0      0 I   1.0   0.0   0:02.17 kworker/u49:13-btrfs-endio                                                                                                    
    11390 root      20   0    9776   4396   2552 R   1.0   0.0   0:00.10 top                                                                                                                           
     2533 root      20   0       0      0      0 I   0.7   0.0   0:20.58 kworker/u49:7-btrfs-endio                                                                                                     
     2621 nobody    20   0 7217200 113880      0 S   0.7   0.1   6:06.13 jackett                                                                                                                       
     7894 root      22   2  113580  24708  19188 S   0.7   0.0   8:01.97 php                                                                                                                           
     9093 root      20   0  283668   3948   3016 S   0.7   0.0   8:15.71 emhttpd                                                                                                                       
    25333 root      20   0 1927136 120228    976 S   0.7   0.1 171:10.80 shfs                                                                                                                          
    31883 nobody    20   0 4400376 491120      4 S   0.7   0.5  15:38.17 Plex Media Serv                                                                                                               
      147 root      20   0       0      0      0 I   0.3   0.0   0:29.49 kworker/14:1-events                                                                                                           
      936 root      20   0  113748  13228   7672 S   0.3   0.0   0:00.22 php-fpm                                                                                                                       
     1662 root      20   0   36104    988      0 D   0.3   0.0   0:07.78 openvpn                                                                                                                       
     1737 root       0 -20       0      0      0 I   0.3   0.0   0:02.95 kworker/12:1H-kblockd                                                                                                         
     2530 root      20   0 8677212 232344      0 S   0.3   0.2   1:09.19 java                                                                                                                          
     2794 nobody    20   0  197352  51372      0 D   0.3   0.1   0:44.64 python                                                                                                                        
     6629 root      20   0       0      0      0 I   0.3   0.0   0:27.33 kworker/8:2-events                                                                                                            
     7607 root      20   0    3656    232    196 D   0.3   0.0   0:00.01 bash                                                                                                                          
    14099 root      20   0   33648  14988     52 D   0.3   0.0   0:13.68 supervisord                                                                                                                   
    18350 root      20   0       0      0      0 I   0.3   0.0   0:42.51 kworker/u50:3-btrfs-endio                                                                                                     
    21018 nobody    20   0 3453756 581936      4 S   0.3   0.6  51:39.60 mono                                                                                                                          
    21121 nobody    20   0 2477100 532884      4 S   0.3   0.5   8:34.30 mono                                                                                                                          
    22859 root      20   0       0      0      0 I   0.3   0.0   0:12.78 kworker/u50:5-btrfs-endio-meta                                                                                                
    25853 root      20   0 2649020  45064  19816 S   0.3   0.0  81:46.26 containerd                                                                                                                    
    31808 root      20   0   76984    664    412 D   0.3   0.0   0:01.23 php7.0                                                                                                                        
    32288 nobody    20   0  429092   1852      0 S   0.3   0.0   0:21.01 Plex Tuner Serv                                                                                                               
    32302 root      20   0       0      0      0 I   0.3   0.0   0:30.08 kworker/19:1-xfs-buf/md9                                                                                                      
        1 root      20   0    2460   1700   1596 S   0.0   0.0   0:13.40 init                                                                                                                          
        2 root      20   0       0      0      0 S   0.0   0.0   0:00.07 kthreadd                                                                                                                      
        3 root       0 -20       0      0      0 I   0.0   0.0   0:00.00 rcu_gp                                                                                                                        
        4 root       0 -20       0      0      0 I   0.0   0.0   0:00.00 rcu_par_gp                                                                                                                    
        6 root       0 -20       0      0      0 I   0.0   0.0   0:00.00 kworker/0:0H-kblockd                                                                                                          
        9 root       0 -20       0      0      0 I   0.0   0.0   0:00.00 mm_percpu_wq                                                                                                                  
       10 root      20   0       0      0      0 S   0.0   0.0   0:27.84 ksoftirqd/0                                                                                                                   
       11 root      20   0       0      0      0 I   0.0   0.0   2:03.96 rcu_sched                                                                                                                     
       12 root      20   0       0      0      0 I   0.0   0.0   0:00.00 rcu_bh                                                                                                                        



    Clearly grafana is having some fun there managing 500% cpu (goes up to like 900% sometimes)


    But even after trying to kill it, it doesn't work, by that I mean it refuses to die, even when force killing entire docker.


    The only thing I can think of recently is that I've started to run some youtube-dl scripts as part of recent yt changes, to archive some channels, but that's hardly doing anything destructive imo, it does write some temp files then remuxes parts into single mkv's etc but that's about it, all done locally by another client as well.


    Attached diagnostics, unraid running atm, but I'll probably kill it before end of the night.


    Any help is appreciated.


  5. 1 hour ago, Squid said:

    What I meant is that you can go from page 1 to 2, and it'll remember what you had checked off and install

    My bad, didn't assume to try it, given it feels compelled to refresh apps on every new view. thanks.


    1 hour ago, jonathanm said:

    He's not talking about the fresh install being that big, it's the ongoing activity where the paths defined INSIDE the container app don't match the paths defined by your docker config. If you have /media mapped to /mnt/user/Media, and you tell your docker app to download to /Media, it's going to fill up the docker image and not end up on the array.

    Ye and I understand that, but again, this hasn't been an issue for ages, I will go over all mounts but all of them are being used correctly from what I could see, since I'd see data dissapearing into .img instead of my libraries and have messed up cases in the past that'd cause this.

  6. 1 hour ago, BRiT said:

    The reason I said to do it one at a time is you obviously have something wrongly mapped on at least one of the dockers. There's no reason to need more space on the actual docker.img. Everything that requires large amounts of space should be mapped through outside the docker container. If you dont review and correct the wrong mapping(s) you will continue to have your original issue.

    Thanks, and ye I get that, after adding fresh templates, no data or anything, it filled 10gb, all dockers should have their own appdata paths for config etc and anything pulling data is using main storage pool as well. I will go over all of them regardless, the only one I can think of would be influxdb which is storing sensors for grafana etc, I believe it should keep its db within appdata as well thought.


    31 minutes ago, Squid said:

    Future reference:  It's not page limited

    What I meant is that I've a number of dockers in use and tested tripple that before as well. so I've 4 pages of old apps, unless im blind or missed some config somewhere I can't display more than 15 items per page under previous apps view. so could do 5-6 at a time (alphabetical) and also again seems to run them in sequence, which is probably intended method of install, but using multiple tabs I could add 5-6 at a time, sure they'd time each other out and cap wan link etc. using apps did "feel" faster though either way, thanks.


    18 minutes ago, trurl said:

    In fact, you should recreate docker image with 20G. Making it larger will just take longer to fill. It won't fix the problem. I've never seen anyone needing more than 20G.


    The usual cause of filling docker image is an application writing to a path that isn't mapped. Your mappings might be correct, but if the application isn't using those mapped paths, it is going to write into the image.


    Typical mistakes are using different upper/lower case when specifying paths, or not using an absolute path (must start with / ).

    I'd agree, and haven't had issues for a long time, like above, after fresh container reinstall from templates, it filled 10gb. I will go over all of them but can't imagine any of them being big enough to need 10gb and even then that'd take like an hour to dl which I'd notice on reinstall.




    what it looks like, ignore disks, need to move some data around to reduce spinups, but ye docker 9.3g fresh installs with no configs

    root@..:~# df -h
    Filesystem      Size  Used Avail Use% Mounted on
    rootfs           48G  762M   47G   2% /
    tmpfs            32M  488K   32M   2% /run
    devtmpfs         48G     0   48G   0% /dev
    tmpfs            48G     0   48G   0% /dev/shm
    cgroup_root     8.0M     0  8.0M   0% /sys/fs/cgroup
    tmpfs           128M   64M   65M  50% /var/log
    /dev/sda1        29G  447M   29G   2% /boot
    /dev/loop0      8.7M  8.7M     0 100% /lib/modules
    /dev/loop1      5.9M  5.9M     0 100% /lib/firmware
    /dev/md1        3.7T  341G  3.4T  10% /mnt/disk1
    /dev/md2        3.7T  841G  2.9T  23% /mnt/disk2
    /dev/md3        3.7T  1.6T  2.2T  43% /mnt/disk3
    /dev/md4        3.7T  232G  3.5T   7% /mnt/disk4
    /dev/md5        7.3T  2.7T  4.7T  37% /mnt/disk5
    /dev/md6        7.3T  3.6T  3.7T  50% /mnt/disk6
    /dev/md7        7.3T  3.5T  3.9T  47% /mnt/disk7
    /dev/md8        7.3T  2.4T  5.0T  33% /mnt/disk8
    /dev/md9        9.1T   11G  9.1T   1% /mnt/disk9
    /dev/md10       932G  290G  642G  32% /mnt/disk10
    /dev/md11       932G  983M  931G   1% /mnt/disk11
    /dev/md12       932G  983M  931G   1% /mnt/disk12
    /dev/md13       932G  983M  931G   1% /mnt/disk13
    /dev/md14       932G  983M  931G   1% /mnt/disk14
    /dev/md15       932G  983M  931G   1% /mnt/disk15
    /dev/md16       932G  983M  931G   1% /mnt/disk16
    /dev/sdz1       466G  103G  363G  23% /mnt/cache
    shfs             60T   16T   44T  26% /mnt/user0
    shfs             60T   16T   45T  26% /mnt/user
    /dev/loop3      1.0G   17M  905M   2% /etc/libvirt
    /dev/loop2       50G  9.3G   41G  19% /var/lib/docker
    shm              64M     0   64M   0% /var/lib/docker/containers/97408ac7186b33a59b3af02cdef729f733f03899aea369a78caee2130a7c9fb4/mounts/shm


  7. 9 hours ago, Squid said:

    Much easier to simply go to the Apps tab, check off everything you want (more than 5 possible), and then hit install Multiple.

    next morning, so restarted docker and its all gone again...  fml.


    So deleted docker.img this time properly and will try adding through apps and see how it goes.



    so adding through apps is more convenient (despite being paged limited) but not faster, running a single docker add at a time compared to 5 tabs etc.



    okay, so manually deleting .img seems to have to worked, after reinstall and docker restart, everything still there. 10gig used after just install of what I acutally use (16, typical plex/sonarr/deluge setup etc). Restoring appdata now and will see what happens.


    plex needs to chill with their file structure;


    1.25m folders and 350k files and still going...

  8. thanks, looks like unraid did that for me, only 500mb used when expanding size, so ended up moving existing data into subdir rather than extract from backup, after going through all templates (seems like unraid only handles 5 adds at once) setting and mappings were preserved which is nice, and now I've stopped docker and I'm moving data back into dir's.


    I'm still assuming that it was the 20gig that filled up, if I didn't force delete the .img will unraid be happy on next boot, don't feel like going through restore again.

  9. Title says it all really.


    so... I suspect it was deluge or plex, not sure, but from the looks of it although .img was full (now expanded to 50gb and 500mb x3 logs). All appdata is there and accessible within cache mirror.


    Anyone know how I can remap the docker containers? profiles still there but don't want to overwrite data on new container creation.


    I have full nightly backups but just extracting 25gig tar would take a night probably.


    any help appreciated.

  10. flooded the library a bit to see what happens, and look like it was on the 2tb pass, filled disk1 until 2tb, and now going for disk 2. With that in mind, I suppose ideally you would keep smaller disks as lower count if you care for "higher use" of high water allocation, as in my case, you won't even touch 4tb drives until you have 18tb's writted to share. I'll expand with 10tb's and re arrange to see what happens, I would expect it to continue filling disk 2 until 2tb free and then look at disks again to choose best target.

  11. 39 minutes ago, itimpi said:

    Unfortunately not - that is the way High Water Allocation works.   It works through drives in slot number order at each switch point.


    If you are really concerned then I think the only solution would be to have the 4TB disks on lower slot numbers than the 8TB ones so they started being used first when the 4TB point was reached.   Since you only have single parity this could be done without recalculating parity if you really wanted to do this.   However is it worth it?

    I see, well thanks for clarifying. Is there a recommended drive allocation when using high water? Also from what you're saying, the high mark threshold/pass is set when it has reached a breakpoint on a single disk? so in theory if I were to put empty drives ahead of 8tb (filled one) when 8tb reaches 2tb high mark, it'd run through disks in order and determine new mark?


    And as you said, its not a huge deal, I'll be upgrading the wd1tb's to 10tb tonight so might move the drives around. So I'll break parity and move it down to data and create a new 10tb parity and 2nd will just go into data. Planning on grabbing a 2nd shelf at some point, but until then will probably rock single parity, given sas shares are sitting empty whilst I prep vmware move.

  12. split set to any at the moment, not bothered about it, might end up setting 3 subdirs in future, but data no big enough to worry about it yet.




    I've moved 900gb to other disk but didn't seem to do much and continue filling up first disk rather than reset to 4tb free mark and fill disk 2 etc.


    I can't confirm but I believe its on the 2tb free pass atm so filling disk 1 until 2tb free mark and then moving down the drives but was hoping to move data around to force it down quicker instead of filling 75% of first disk before even touching empty 4tb drives

  13. Hey,


    So a quick one hopefully, so recently added more disks to my library share, which are lower volume than original and current disks, i.e had 3x8tb and now added 4x4tb.


    Share is in high water allocation, I was wondering if running unbalance scatter job to rebalance data across the 7 drives would reset the high water "pass" count and instead of continuing on its current (2tb free atm) pass and retry from the start (i.e. 4tb free) given reduced disk space? I hope that makes sense? 

  14. Big thanks to you and  your team, overall solid product and happy to continue using it.




    Do you plan to develop the hypervisor functionality further? such as including common gpu hardware drivers like nvidia, amd or intel for VM passthrough, which is currently achieved through community projects.




    SSD's are currently experimental and not fully supported, do you plan on bringing them along with nvme drives as future data mediums change (also ties into q1 in terms of drivers)?




    If you were to start the whole project again where would you focus your time or ditch some features/projects altogether?

  15. just reviving because 1st result in google;


    binhex sonarr template maps /data to its appdata and it's what binhex deluge/vpn uses as its media path. so you'll need to map a path that both can access/use/rw and match in both mapping and target path, i.e. /completed/ to /mnt/user0/data/completed.


    For proper automation, jackett grabs indexes, sonarr grabs all files, picks best match and throws it over to deluge with label of whatever set to move to dl to /downloads/ which sonarr has mapped as well and is able to pick up that file, then hardlink/copy and rename in right path that your media player can pick up in correct and renamed format. Then you just set label to seed to whatever, forever if you care, or just seed low value like .1 and delete on target which will remove badly named seed and it'll keep hardlink alive and well in right place.


    Works well, with few rss feeds to auto feed new titles in etc which sonarr will add and pick up future releases, can just leave it to do its thing.

  16. So started weekend project to migrate from freenas to unraid, all looking good so far, apart from expected 1.6days for data transfer and 40days for parity build lmao


    Anyway, moving from vm hosted yassr2 deluge.


    So turns out you can just move your old configs and eggs from (windows) %appdata%\deluge and drop them into your config path, default being /mnt/user/appdata/deluge..


    although my plugins path is super broken pointing to c:\..\appdata\ it still picks up plugins dir fine and installs them correctly and enables according to .conf


    And since im such a nice guy, here's my mega links if you want something to go off of with no previous setup, my current is just looking at some nyaa rss, so adapt as you need, very easy to read and make changes.






    after enabling in web (will reset to off and report in deluged-web.log it has no ui)

    [INFO    ] 10:58:58 pluginmanager:108 'YaRSS2' plugin contains no WebUI code, ignoring WebUI enable call.
    [INFO    ] 10:58:59 pluginmanager:145 Plugin has no web ui


    but output from deluged.log


    [INFO    ] 10:44:07 rpcserver:206 Deluge Client connection made from:
    [INFO    ] 10:44:41 logger:37 YaRSS2.__init__:34: Appending to sys.path: '/config/plugins/YaRSS2-1.4.1-py2.7.egg/yarss2/include'
    [INFO    ] 10:44:41 logger:37 YaRSS2.rssfeed_scheduler:42: Scheduled RSS Feed '1080p HorribleSubs Nyaa' with interval 5
    [INFO    ] 10:44:41 logger:37 YaRSS2.rssfeed_scheduler:42: Scheduled RSS Feed '1080p BlueLobster Nyaa' with interval 5
    [INFO    ] 10:44:41 logger:37 YaRSS2.core:40: Enabled YaRSS2 1.4.1
    [INFO    ] 10:44:41 pluginmanagerbase:158 Plugin YaRSS2 enabled..




    INFO    ] 11:03:59 logger:37 YaRSS2.rssfeed_handling:267: Update handler executed on RSS Feed '1080p BlueLobster Nyaa (nyaa.si)' (Update interval 5 min)
    [INFO    ] 11:03:59 logger:37 YaRSS2.rssfeed_handling:310: Fetching subscription '1080p BlueLobster Nyaa'.
    [INFO    ] 11:03:59 logger:37 YaRSS2.rssfeed_handling:80: Fetching RSS Feed: '1080p BlueLobster Nyaa' with Cookie: '{}' and User-agent: 'Deluge v1.3.15 YaRSS2 v1.4.1 Linux/4.19.41-Unraid'.
    [INFO    ] 11:04:00 logger:37 YaRSS2.rssfeed_handling:332: 70 items in feed, 70 matches the filter.
    [INFO    ] 11:04:00 logger:37 YaRSS2.rssfeed_handling:344: Not adding because of old timestamp: '[BlueLobster] Meow Meow Japanese History (01-96) [1080p] (Batch) (Neko Neko Nihonshi)'



    and proceed to use yarss2.conf file as configured.


    added 2nd sub, just needs a "1" :{..} and another for feed, tested and works fine, make sure they're unique and match against right feeds


    like so;