Dimtar

Members
  • Posts

    549
  • Joined

  • Last visited

Posts posted by Dimtar

  1. 2 minutes ago, sparklyballs said:

     

    not that i'm aware of i don't use plex let alone tautulli but just going through the motions of adding it i can change the host port 

    you don't want to ever change the container port as a rule anyways

     

     

    For comparison this is the Sonnar container:

    image.png.f7e5de24ccd8d4b2e3004d090cb3fb5c.png

     

    I obviously don't know anything about these things but my guess is the template isn't aware the Tautulli container needs a port?

  2. 3 minutes ago, sparklyballs said:

     

    we don't recommend updating through the webui

     

    for instance you'd be missing a dependency that was added as part of the rebrand 

     

    Thanks Sparky, normally I don't update through the container but a broken container wouldn't worry me in this case.

     

    Is there something wrong with the tautilli container template? It seems like its missing the container port?

    image.png.3046707f6252d30f66c92dc655c57df3.png

     

    image.png.92b4e6e151d93eefcfc5dcc120f74db9.png

  3. 9 hours ago, CHBMB said:

    Project has been renamed from Plexpy to Tautulli.

     

    To migrate:

     

    1.  Stop the current plexpy container.

    2.  Use Community Applications to install Tautulli, utilising your existing Plexpy appdata folder for /config.

     

    I clicked the update prompt in the web interface for PlexPy and updated to Tautulli without hassle. It also works after a restart of the container. Are you saying the plexpy and tautulli containers are different?

  4. 2 hours ago, ezbox said:

    Good Afternoon,

     

    It seems the latest update of the docker is broken, does not load up, had to go to version 91 again and this fixed the problem.

     

    if you need to do this, change the repository to linuxserver/radarr:91

     

    Just reporting,

     

    Ezbox

     

    PS. Another friend who uses the docker too had the same problem, unable to load.

     

    Just wanted throw my hat in the ring on this one.

  5. 6 hours ago, WannabeMKII said:

    I'm trying to migrate from v1 to v2. As per the instructions, I've moved the cfg and db files to the root of the hydra2, but when trying to import, I get the following;

     

    But as per the screenshot, the files are there and in the right place?

     

    Where have I gone wrong?

     

     

     

     

    Try /config/settings.cfg /config/nzbhydra.db

    • Upvote 1
  6. On 16/01/2018 at 10:06 AM, saarg said:

    I guess this will be the one as there may be some issues migrating to V2. Don't want V1 users to suddenly have a non working container if we pushed V2.

     It’s a major new version running on an entirely different language, I think this is the right approach.

     

    On another note, the container is great. Thanks team.

  7. Here is a very rough example.

    rclone mount --cache-db-path /mnt/user/rclonedb --cache-chunk-path /mnt/user/rclonechunk googlecache: /mnt/user/googlemount

    You can just set it all in line, the above example would only work if you created the shares before hand and hopefully put them on a cache drive. Hope this helps somewhat.

  8. On 21/01/2018 at 12:01 AM, johnnie.black said:

    You're getting out of memory errors, probably one of the dockers, best bet would be to shut them all down, turn on one by one and leave it a few days to try and find the culprit.

     

    Why am I not getting email notifications of your replies? This comment is not directed at you, just speaking out loud.

     

    Thanks for the tip, I am pretty sure its the NZBGet docker as its the only container doing anything each day. I also install iotop to help me, thanks for your help I appreciate it.

  9. Hi all.

     

    This problem wasn't happening but then it started randomly and its happening more and more. So right now, my server is near un-usable. Its currently sitting at a load of 60. (please see screenshot)

    The server isn't doing anything, the load will at some stage come down and the server will be usable again but I am trying to work out what exactly is causing this. If i open netdata during this time, it shows IOwait as the biggest issue but I am not sure what is causing that. 

     

    This is the result of "free -hm", am I running out of ram and its swapping to drives?


    root@ASTRO:~# free -hm
                  total        used        free      shared  buff/cache   available
    Mem:           3.6G        2.9G        190M        473M        558M         14M
    Swap:            0B          0B          0B

     

    Like right now I have 10% of kswap0 and 10% of unraidd, is that it swapping to the disk?

     

    I don't have a cache drive and the server is essentially ideal right now. Any ideas on stuff to look at?

     

    putty_2018-01-19_22-59-37.png

  10. 11 hours ago, ScoHo said:

    Trying to set this up with Google Drive and getting the below message when the backup runs. I did enable no-auto-compact. I'm running a fresh backup for the first time so I don't care about losing data on the Google Drive (there isn't any there anyway). Any ideas?

     

    "The backup contains files that belong to another operating system. Proceeding with a backup would cause the database to contain paths from two different operation systems, which is not supported. To proceed without losing remote data, delete all filesets and make sure the --no-auto-compact option is set, then run the backup again to re-use the existing data on the remote store."

     

    Could you try backing up somewhere else? Like B2, just as a test don't do the full backup. Just to work out if its Google Drive or something else?

  11. Hi all.

     

    So there has been alot of videos and talk about having an unRaid box, running a VM and passing through the graphics card. It makes me wonder, who is actually using this day to day? As in they have a fully functioning desktop PC capable of running Windows with a keyboard/mouse/monitor and running unRaid underneath. Keen to hear all stories.

  12. 20 minutes ago, binhex said:

     

    11GB, impressive, it must be logging the entire internet :-), i will take a look at this when i get a min, see if i can at the very least redirect it outside of the container.

     

    I have hardly even used it, maybe its just really noisy? Thanks for looking into this, whenever you do.

  13. 3 minutes ago, DZMM said:

    Unfortunately you have to delete your docker image and reinstall each docker (you don't lose app data) afterwards.

     

    Thanks for the fast reply. I added the extra parameters you mentioned and this was enough to give the container a new ID. This deleted the old folder and removed the 11GB file its self. So all should be good now, thanks again.