• Inconsistent but frequent NIC errors NIC Link is Down / Up


    Marshalleq
    • Minor

    Apr 23 16:34:01 Obi-Wan kernel: mdcmd (248): spindown 0

    Apr 23 16:34:03 Obi-Wan kernel: mdcmd (249): spindown 6

    Apr 23 16:36:34 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Down

    Apr 23 16:36:34 Obi-Wan kernel: br0: port 1(eth0) entered disabled state

    Apr 23 16:36:36 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: RX

    Apr 23 16:36:36 Obi-Wan kernel: br0: port 1(eth0) entered blocking state

    Apr 23 16:36:36 Obi-Wan kernel: br0: port 1(eth0) entered forwarding state

     

    I just got another one while typing this (am just doing a live tail on the syslog) - this one appears to be more about docker I think being a bridge.  I am actually now unsure if these are normal ie part of just a docker restart or if the interface is actually restarting for the whole system?

     

    Apr 23 17:03:10 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Down

    Apr 23 17:03:10 Obi-Wan kernel: br0: port 1(eth0) entered disabled state

    Apr 23 17:03:12 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: RX

    Apr 23 17:03:13 Obi-Wan ntpd[2659]: Deleting interface #58 br0, 192.168.43.10#123, interface stats: received=12, sent=12, dropped=0, active_time=5880 secs

    Apr 23 17:03:13 Obi-Wan ntpd[2659]: 192.168.43.3 local addr 192.168.43.10 -> <null>

    Apr 23 17:03:13 Obi-Wan ntpd[2659]: Deleting interface #59 br0, fe80::24d8:79ff:fe01:a2fc%9#123, interface stats: received=0, sent=0, dropped=0, active_time=5880 secs

    Apr 23 17:03:13 Obi-Wan kernel: br0: port 1(eth0) entered blocking state

    Apr 23 17:03:13 Obi-Wan kernel: br0: port 1(eth0) entered forwarding state

    Apr 23 17:03:15 Obi-Wan ntpd[2659]: Listen normally on 60 br0 192.168.43.10:123

    Apr 23 17:03:15 Obi-Wan ntpd[2659]: Listen normally on 61 br0 [fe80::24d8:79ff:fe01:a2fc%9]:123

    Apr 23 17:03:15 Obi-Wan ntpd[2659]: new interface(s) found: waking up resolver

     

    Diagnostics attached.

    obi-wan-diagnostics-20190423-0509.zip




    User Feedback

    Recommended Comments

    So thinking through if this is normal or not and how to test it, I ran a ping with a timestamp.

     

    First the error (yet another one), then the ping.

     

    Apr 23 17:12:13 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Down

    Apr 23 17:12:13 Obi-Wan kernel: br0: port 1(eth0) entered disabled state

    Apr 23 17:12:15 Obi-Wan kernel: igb 0000:05:00.0 eth0: igb: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: RX

    Apr 23 17:12:15 Obi-Wan kernel: br0: port 1(eth0) entered blocking state

    Apr 23 17:12:15 Obi-Wan kernel: br0: port 1(eth0) entered forwarding state

     

    Tue Apr 23 17:12:09 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=191 ttl=64 time=0.472 ms

    Tue Apr 23 17:12:10 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=192 ttl=64 time=0.440 ms

    Tue Apr 23 17:12:11 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=193 ttl=64 time=0.176 ms

    Tue Apr 23 17:12:12 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=194 ttl=64 time=0.194 ms

    Tue Apr 23 17:12:16 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=198 ttl=64 time=0.144 ms

    Tue Apr 23 17:12:17 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=199 ttl=64 time=0.274 ms

    Tue Apr 23 17:12:18 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=200 ttl=64 time=0.270 ms

    Tue Apr 23 17:12:19 NZST 2019: 64 bytes from 192.168.43.3: icmp_seq=201 ttl=64 time=0.416 ms

     

    You can see, while the NIC was down at 17:12:13 to 17:12:15, pings stopped (see 17:12:12 then 17:12:16.

     

    Coincidence?  I don't think so.  Whether this makes much of a difference to anything I don't know, but I could see in an online game how having the network down for four whole seconds might wreak some havoc.

    Link to comment

    I've now powered down and up the switch (which didn't work), and have now just tried enabling flow control on the switch port - probably not working but because Unraid seems to have pause on in RX mode.  Tomorrow will try another cable and another NIC.  I'm thinking it's software though.

    Link to comment

    "link is down" is more likely a hardware issue.

    • Check or replace cables
    • Use a different NIC
    • Use a different switch port
    • Use a different switch
    Link to comment

    I've already checked cables, changed the switch port, I still can't try a new NIC because CrashPlan and it's block sync is on it's third day sigh and stopping it would be beginning from the beginning!  I think it will be finished tomorrow.  I have my doubts about the switch being an issue, I can check on another linux box if it is having a similar issue, but I doubt it.  Thanks for your input, will try card next.

    Link to comment

    I have something similar in my logs, it's been going on under 6.6.7 prior to me updating to 6.7.0-rc7. It happens several times a day. I recently switched both the switch that Unraid is connected to and the ethernet card, but I'm seeing similar messages both before and after I switched the hardware.

     

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered blocking state

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered disabled state

    Apr 23 10:42:24 unRAID kernel: device veth36c9b86 entered promiscuous mode

    Apr 23 10:42:24 unRAID kernel: IPv6: ADDRCONF(NETDEV_UP): veth36c9b86: link is not ready

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered blocking state

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered forwarding state

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered disabled state

    Apr 23 10:42:24 unRAID kernel: eth0: renamed from vethaf1be8e

    Apr 23 10:42:24 unRAID kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth36c9b86: link becomes ready

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered blocking state

    Apr 23 10:42:24 unRAID kernel: docker0: port 5(veth36c9b86) entered forwarding state

    Apr 23 10:42:27 unRAID kernel: docker0: port 6(vethe876c38) entered blocking state

    Apr 23 10:42:27 unRAID kernel: docker0: port 6(vethe876c38) entered disabled state

    Apr 23 10:42:27 unRAID kernel: device vethe876c38 entered promiscuous mode

    Apr 23 10:42:27 unRAID kernel: IPv6: ADDRCONF(NETDEV_UP): vethe876c38: link is not ready

    Apr 23 10:42:27 unRAID kernel: docker0: port 6(vethe876c38) entered blocking state

    Apr 23 10:42:27 unRAID kernel: docker0: port 6(vethe876c38) entered forwarding state

    Apr 23 10:42:27 unRAID kernel: docker0: port 6(vethe876c38) entered disabled state

     

    Apr 23 14:57:29 unRAID kernel: br0: port 3(vnet1) entered blocking state

    Apr 23 14:57:29 unRAID kernel: br0: port 3(vnet1) entered disabled state

    Apr 23 14:57:29 unRAID kernel: device vnet1 entered promiscuous mode

    Apr 23 14:57:29 unRAID kernel: br0: port 3(vnet1) entered blocking state

    Apr 23 14:57:29 unRAID kernel: br0: port 3(vnet1) entered forwarding state

     

    I was going to get around to asking the question at some point if these messages matter. I have not had any noticeable performance or network issues so it hasn't been a priority.

     

     

     

     

    Link to comment
    2 hours ago, italeffect said:

    I was going to get around to asking the question at some point if these messages matter. I have not had any noticeable performance or network issues so it hasn't been a priority.

    These messages come from Docker, harmless and expected.

    Link to comment

    Thanks.  Yes, I think those are different as with yours there is not a primary network connection going down and up, whereas with mine, I think it is.

    Link to comment
    6 hours ago, Marshalleq said:

    I've already checked cables, changed the switch port, I still can't try a new NIC because CrashPlan and it's block sync is on it's third day sigh and stopping it would be beginning from the beginning!  I think it will be finished tomorrow.  I have my doubts about the switch being an issue, I can check on another linux box if it is having a similar issue, but I doubt it.  Thanks for your input, will try card next.

    Reseating (or moving) the NIC may help too.

    Intel NICs are the most commonly used in Unraid and work well.

    Link to comment


    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Restore formatting

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Status Definitions

     

    Open = Under consideration.

     

    Solved = The issue has been resolved.

     

    Solved version = The issue has been resolved in the indicated release version.

     

    Closed = Feedback or opinion better posted on our forum for discussion. Also for reports we cannot reproduce or need more information. In this case just add a comment and we will review it again.

     

    Retest = Please retest in latest release.


    Priority Definitions

     

    Minor = Something not working correctly.

     

    Urgent = Server crash, data loss, or other showstopper.

     

    Annoyance = Doesn't affect functionality but should be fixed.

     

    Other = Announcement or other non-issue.