• [6.6.0-rc4] Docker Containers Being Randomly Deleted


    xthursdayx
    • Retest Annoyance

    I noticed today that my Letsencrypt and Jackett containers were removed for no reason that I can find. Their orphaned images were still visible in my docker, however. After looking at my System Log all I can see is that perhaps the Community Apps autoupdate failed? What do you all think?

     

    Here is the (seemingly) relevant log section:

    Nov 20 05:07:53 vulfTower Docker Auto Update: Stopping jackett
    Nov 20 05:07:53 vulfTower Docker Auto Update: docker stop -t 10 jackett
    Nov 20 05:08:33 vulfTower kernel: veth89c2e86: renamed from eth0
    Nov 20 05:08:33 vulfTower kernel: docker0: port 7(vetha3dfdd5) entered disabled state
    Nov 20 05:08:51 vulfTower kernel: docker0: port 7(vetha3dfdd5) entered disabled state
    Nov 20 05:08:51 vulfTower avahi-daemon[12509]: Interface vetha3dfdd5.IPv6 no longer relevant for mDNS.
    Nov 20 05:08:51 vulfTower avahi-daemon[12509]: Leaving mDNS multicast group on interface vetha3dfdd5.IPv6 with address fe80::70c2:94ff:fe16:95f0.
    Nov 20 05:08:51 vulfTower kernel: device vetha3dfdd5 left promiscuous mode
    Nov 20 05:08:51 vulfTower kernel: docker0: port 7(vetha3dfdd5) entered disabled state
    Nov 20 05:08:51 vulfTower avahi-daemon[12509]: Withdrawing address record for fe80::70c2:94ff:fe16:95f0 on vetha3dfdd5.
    Nov 20 05:09:24 vulfTower Docker Auto Update: Stopping letsencrypt
    Nov 20 05:09:24 vulfTower Docker Auto Update: docker stop -t 10 letsencrypt
    Nov 20 05:09:43 vulfTower kernel: vetheb320f2: renamed from eth0
    Nov 20 05:09:43 vulfTower kernel: docker0: port 9(vethbff63ff) entered disabled state
    Nov 20 05:10:00 vulfTower avahi-daemon[12509]: Interface vethbff63ff.IPv6 no longer relevant for mDNS.
    Nov 20 05:10:00 vulfTower avahi-daemon[12509]: Leaving mDNS multicast group on interface vethbff63ff.IPv6 with address fe80::8863:5cff:fe40:6834.
    Nov 20 05:10:00 vulfTower kernel: docker0: port 9(vethbff63ff) entered disabled state
    Nov 20 05:10:00 vulfTower kernel: device vethbff63ff left promiscuous mode
    Nov 20 05:10:00 vulfTower kernel: docker0: port 9(vethbff63ff) entered disabled state
    Nov 20 05:10:00 vulfTower avahi-daemon[12509]: Withdrawing address record for fe80::8863:5cff:fe40:6834 on vethbff63ff.
    Nov 20 05:10:36 vulfTower Docker Auto Update: Installing Updates for jackett letsencrypt
    Nov 20 05:23:29 vulfTower Docker Auto Update: Restarting jackett
    Nov 20 05:23:30 vulfTower Docker Auto Update: Restarting letsencrypt
    Nov 20 05:23:30 vulfTower sSMTP[24538]: Creating SSL connection to host
    Nov 20 05:23:30 vulfTower sSMTP[24538]: SSL connection using ECDHE-RSA-AES256-GCM-SHA384
    Nov 20 05:23:30 vulfTower sSMTP[24538]: Authorization failed (535 5.7.1 Authentication failed)
    Nov 20 05:23:30 vulfTower Docker Auto Update: Community Applications Docker Autoupdate finished

     




    User Feedback

    Recommended Comments

    Great, I didn't realize that this newer version was available. I've upgraded and everything looks okay now (after having reinstalled those two docker containers). I'll update if anything changes.

    Link to comment

    So, I spoke too soon. This evening I saw that LetsEncrypt had an update available, so I clicked "update container" in the docker menu. This process stalled without completing while "removing the image" and I eventually had to reload a new docker page and delete the orphaned image. I then went to the Community Apps section in order to install LetsEncrypt again only to find that it stalls halfway through the install at around the "Pulling fs layer" part of the process. However, I opened a dashboard page and found that the container did finish installing (though the original page never showed the install completing). But, strangely enough, the container show's as having an available update... I'm not installing this update, so as to avoid chasing my own tail, but I suspect that it would lead to the same situation described above. 

     

    Here is the (seemingly) relevant log section:

    Nov 21 00:52:09 vulfTower nginx: 2018/11/21 00:52:09 [error] 12502#12502: *89462 upstream timed out (110: Connection timed out) while reading upstream, client: 192.168.1.1, server: , request: "POST /Apps/UpdateContainer?xmlTemplate=user:/boot/config/plugins/dockerMan/templates-user/my-letsencrypt.xml HTTP/1.1", upstream: "fastcgi://unix:/var/run/php5-fpm.sock:", host: "192.168.1.107", referrer: "http://192.168.1.107/Apps/UpdateContainer?xmlTemplate=user:/boot/config/plugins/dockerMan/templates-user/my-letsencrypt.xml"

     

    Any ideas?

    Edited by zandrsn
    Link to comment

    Start with posting your diagnostics file. See Tools -> Diagnostics

     

    As a quick test start your system in safemode and see if the problem persists.

    Link to comment
    On 11/21/2018 at 1:31 AM, bonienl said:

    Start with posting your diagnostics file. See Tools -> Diagnostics

     

    As a quick test start your system in safemode and see if the problem persists.

    Sorry for the delay, I was away from my server. After upgrading to 6.6.5 auto-updating seems to work fine, but I do find that whenever I try to update containers manually it usually stalls, resulting in an orphaned image and need to reinstall the container app, and when I try to stop or restart a container I usually get an "Execution error" message and have to reload the docker page to see if the container actually stopped or restarted.

     

    I've attached my diagnostics below:

    vulftower-diagnostics-20181127-0926.zip

    Edited by zandrsn
    clarified container restarting problem
    Link to comment

    I'm also getting this in my system log. It may be totally unrelated, so apologies if so, but I thought it was strange because I don't have a drive called sdk1
     

    Nov 27 09:27:46 vulfTower kernel: fat__get_entry: 6 callbacks suppressed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8364) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8365) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8366) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8367) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8368) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8369) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8370) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8371) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8372) failed
    Nov 27 09:27:46 vulfTower kernel: FAT-fs (sdk1): Directory bread(block 8373) failed

     

    Edited by zandrsn
    added code
    Link to comment
    17 minutes ago, zandrsn said:

    FAT-fs

    Your flashdrive most likely dropped offline.  Use a different port (preferably USB2).  You will have to reboot.

    Link to comment
    6 minutes ago, Squid said:

    Your flashdrive most likely dropped offline. 

    1

     

    That's what I thought as well, because of the FAT-fs, but it's still showing as mounted at sda.

    My server is running a parity check right now, but I'll reboot afterward and see if the error persists.
     

    Edited by zandrsn
    spelling
    Link to comment

    Sure, make me actually read the diagnostics :(

     

    Nov 25 23:55:05 vulfTower kernel: scsi 18:0:0:0: Direct-Access     Kindle   Internal Storage 0100 PQ: 0 ANSI: 2
    Nov 25 23:55:05 vulfTower kernel: sd 18:0:0:0: Attached scsi generic sg11 type 0
    Nov 25 23:55:05 vulfTower kernel: sd 18:0:0:0: Power-on or device reset occurred
    Nov 25 23:55:05 vulfTower kernel: sd 18:0:0:0: [sdk] Attached SCSI removable disk

     

    its your kindle 

    Link to comment
    10 hours ago, Squid said:

    its your kindle 

     

    Weird... there wasn't a kindle attached. Maybe some sort of lingering mount point from when one was previously charging off of the server. 

     

    Thanks for checking out the diagnostics. Did you happen to see anything that might indicate why docker container manual updating is failing (and orphaning images) so frequently, and restarting or stopping so often results in an execution error?

    Link to comment

    ... it might be related to a (the) ssl bug i saw here somewhere? 

     

    besides this i also notcied that sometimes manual updating / installing dockers result in a hanging webpage until it shows in the log upstream timeout... but never happend in auto update *knock on wood* but since unraid istnw orking since i updated to the newest rc i cant tell for sure.

     

    edit there it is:

     

     

    Edited by nuhll
    Link to comment


    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Restore formatting

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Status Definitions

     

    Open = Under consideration.

     

    Solved = The issue has been resolved.

     

    Solved version = The issue has been resolved in the indicated release version.

     

    Closed = Feedback or opinion better posted on our forum for discussion. Also for reports we cannot reproduce or need more information. In this case just add a comment and we will review it again.

     

    Retest = Please retest in latest release.


    Priority Definitions

     

    Minor = Something not working correctly.

     

    Urgent = Server crash, data loss, or other showstopper.

     

    Annoyance = Doesn't affect functionality but should be fixed.

     

    Other = Announcement or other non-issue.