Problems with bonding on 16.04 LTS

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP








up vote
1
down vote

favorite












The system runs Ubuntu 16.04.3 LTS with 3 1GBit NICs: one embedded and 2 Intel PCIe NICs. Both Intel NICs are bonded (bond0) with mode 4 (LACP). The switch is confugured to support LACP on these 2 ports. Here is the network configuration:



cat /etc/network/interfaces
# This file describes the network interfaces available on your system
# and how to activate them. For more information, see interfaces(5).

source /etc/network/interfaces.d/*

# The loopback network interface
auto lo
iface lo inet loopback

# The primary network interface
auto enp0s31f6
#iface enp0s31f6 inet dhcp
iface enp0s31f6 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
gateway 192.168.x.1
dns-nameservers 192.168.x.x

auto enp3s0
iface enp3s0 inet manual
bond-master bond0

auto enp4s0
iface enp4s0 inet manual
bond-master bond0

auto bond0
iface bond0 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
bond-mode 802.3ad
bond-miimon 100
bond-lacp-rate 1
bond-slaves none


This configuration runs rather well without errors. But if the network load is rather high (for instance, during copying 100-200 GB), the following errors are produced in /var/log/syslog:



Feb 14 17:20:02 ubuntu1 kernel: [29601.287684] e1000e: enp3s0 NIC Link is Down
Feb 14 17:20:02 ubuntu1 kernel: [29601.287993] e1000e 0000:03:00.0 enp3s0: speed changed to 0 for port enp3s0
Feb 14 17:20:02 ubuntu1 kernel: [29601.379193] bond0: link status definitely down for interface enp3s0, disabling it
Feb 14 17:20:02 ubuntu1 kernel: [29601.379199] bond0: first active interface up!
Feb 14 17:20:04 ubuntu1 kernel: [29603.064712] e1000e: enp3s0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: Rx/Tx
Feb 14 17:20:04 ubuntu1 kernel: [29603.079162] bond0: link status definitely up for interface enp3s0, 1000 Mbps full duplex


Is it a known problem? Apparently after some seconds the failed interface works well again. The problem doesn't occur very often.



In the file /proc/net/bonding/bond0 I can see, that the mode 4 was correctly recognized:



cat /proc/net/bonding/bond0 
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)

Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2 (0)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

802.3ad info
LACP rate: fast
Min links: 0
Aggregator selection policy (ad_select): stable
System priority: 65535


I have tried to use bond-slaves with interface names instead of none. But in this case the ifenslave was locked during restarting of network services. So I have found a recommendation, that using "none" the bond0 will be up and doesn't lock.



Any ideas?










share|improve this question























  • Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
    – Terrance
    Feb 14 at 17:00










  • Do you have any suggestions how to trace or diagnostic this problem?
    – Andrej
    Feb 14 at 17:20










  • Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
    – Terrance
    Feb 14 at 17:22










  • This is Linksys LGS326 26-Port Gigabit Smart Switch.
    – Andrej
    Feb 14 at 17:28










  • I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
    – Terrance
    Feb 14 at 18:05














up vote
1
down vote

favorite












The system runs Ubuntu 16.04.3 LTS with 3 1GBit NICs: one embedded and 2 Intel PCIe NICs. Both Intel NICs are bonded (bond0) with mode 4 (LACP). The switch is confugured to support LACP on these 2 ports. Here is the network configuration:



cat /etc/network/interfaces
# This file describes the network interfaces available on your system
# and how to activate them. For more information, see interfaces(5).

source /etc/network/interfaces.d/*

# The loopback network interface
auto lo
iface lo inet loopback

# The primary network interface
auto enp0s31f6
#iface enp0s31f6 inet dhcp
iface enp0s31f6 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
gateway 192.168.x.1
dns-nameservers 192.168.x.x

auto enp3s0
iface enp3s0 inet manual
bond-master bond0

auto enp4s0
iface enp4s0 inet manual
bond-master bond0

auto bond0
iface bond0 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
bond-mode 802.3ad
bond-miimon 100
bond-lacp-rate 1
bond-slaves none


This configuration runs rather well without errors. But if the network load is rather high (for instance, during copying 100-200 GB), the following errors are produced in /var/log/syslog:



Feb 14 17:20:02 ubuntu1 kernel: [29601.287684] e1000e: enp3s0 NIC Link is Down
Feb 14 17:20:02 ubuntu1 kernel: [29601.287993] e1000e 0000:03:00.0 enp3s0: speed changed to 0 for port enp3s0
Feb 14 17:20:02 ubuntu1 kernel: [29601.379193] bond0: link status definitely down for interface enp3s0, disabling it
Feb 14 17:20:02 ubuntu1 kernel: [29601.379199] bond0: first active interface up!
Feb 14 17:20:04 ubuntu1 kernel: [29603.064712] e1000e: enp3s0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: Rx/Tx
Feb 14 17:20:04 ubuntu1 kernel: [29603.079162] bond0: link status definitely up for interface enp3s0, 1000 Mbps full duplex


Is it a known problem? Apparently after some seconds the failed interface works well again. The problem doesn't occur very often.



In the file /proc/net/bonding/bond0 I can see, that the mode 4 was correctly recognized:



cat /proc/net/bonding/bond0 
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)

Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2 (0)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

802.3ad info
LACP rate: fast
Min links: 0
Aggregator selection policy (ad_select): stable
System priority: 65535


I have tried to use bond-slaves with interface names instead of none. But in this case the ifenslave was locked during restarting of network services. So I have found a recommendation, that using "none" the bond0 will be up and doesn't lock.



Any ideas?










share|improve this question























  • Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
    – Terrance
    Feb 14 at 17:00










  • Do you have any suggestions how to trace or diagnostic this problem?
    – Andrej
    Feb 14 at 17:20










  • Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
    – Terrance
    Feb 14 at 17:22










  • This is Linksys LGS326 26-Port Gigabit Smart Switch.
    – Andrej
    Feb 14 at 17:28










  • I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
    – Terrance
    Feb 14 at 18:05












up vote
1
down vote

favorite









up vote
1
down vote

favorite











The system runs Ubuntu 16.04.3 LTS with 3 1GBit NICs: one embedded and 2 Intel PCIe NICs. Both Intel NICs are bonded (bond0) with mode 4 (LACP). The switch is confugured to support LACP on these 2 ports. Here is the network configuration:



cat /etc/network/interfaces
# This file describes the network interfaces available on your system
# and how to activate them. For more information, see interfaces(5).

source /etc/network/interfaces.d/*

# The loopback network interface
auto lo
iface lo inet loopback

# The primary network interface
auto enp0s31f6
#iface enp0s31f6 inet dhcp
iface enp0s31f6 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
gateway 192.168.x.1
dns-nameservers 192.168.x.x

auto enp3s0
iface enp3s0 inet manual
bond-master bond0

auto enp4s0
iface enp4s0 inet manual
bond-master bond0

auto bond0
iface bond0 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
bond-mode 802.3ad
bond-miimon 100
bond-lacp-rate 1
bond-slaves none


This configuration runs rather well without errors. But if the network load is rather high (for instance, during copying 100-200 GB), the following errors are produced in /var/log/syslog:



Feb 14 17:20:02 ubuntu1 kernel: [29601.287684] e1000e: enp3s0 NIC Link is Down
Feb 14 17:20:02 ubuntu1 kernel: [29601.287993] e1000e 0000:03:00.0 enp3s0: speed changed to 0 for port enp3s0
Feb 14 17:20:02 ubuntu1 kernel: [29601.379193] bond0: link status definitely down for interface enp3s0, disabling it
Feb 14 17:20:02 ubuntu1 kernel: [29601.379199] bond0: first active interface up!
Feb 14 17:20:04 ubuntu1 kernel: [29603.064712] e1000e: enp3s0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: Rx/Tx
Feb 14 17:20:04 ubuntu1 kernel: [29603.079162] bond0: link status definitely up for interface enp3s0, 1000 Mbps full duplex


Is it a known problem? Apparently after some seconds the failed interface works well again. The problem doesn't occur very often.



In the file /proc/net/bonding/bond0 I can see, that the mode 4 was correctly recognized:



cat /proc/net/bonding/bond0 
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)

Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2 (0)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

802.3ad info
LACP rate: fast
Min links: 0
Aggregator selection policy (ad_select): stable
System priority: 65535


I have tried to use bond-slaves with interface names instead of none. But in this case the ifenslave was locked during restarting of network services. So I have found a recommendation, that using "none" the bond0 will be up and doesn't lock.



Any ideas?










share|improve this question















The system runs Ubuntu 16.04.3 LTS with 3 1GBit NICs: one embedded and 2 Intel PCIe NICs. Both Intel NICs are bonded (bond0) with mode 4 (LACP). The switch is confugured to support LACP on these 2 ports. Here is the network configuration:



cat /etc/network/interfaces
# This file describes the network interfaces available on your system
# and how to activate them. For more information, see interfaces(5).

source /etc/network/interfaces.d/*

# The loopback network interface
auto lo
iface lo inet loopback

# The primary network interface
auto enp0s31f6
#iface enp0s31f6 inet dhcp
iface enp0s31f6 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
gateway 192.168.x.1
dns-nameservers 192.168.x.x

auto enp3s0
iface enp3s0 inet manual
bond-master bond0

auto enp4s0
iface enp4s0 inet manual
bond-master bond0

auto bond0
iface bond0 inet static
mtu 9000
address 192.168.x.x
netmask 255.255.x.0
network 192.168.x.0
bond-mode 802.3ad
bond-miimon 100
bond-lacp-rate 1
bond-slaves none


This configuration runs rather well without errors. But if the network load is rather high (for instance, during copying 100-200 GB), the following errors are produced in /var/log/syslog:



Feb 14 17:20:02 ubuntu1 kernel: [29601.287684] e1000e: enp3s0 NIC Link is Down
Feb 14 17:20:02 ubuntu1 kernel: [29601.287993] e1000e 0000:03:00.0 enp3s0: speed changed to 0 for port enp3s0
Feb 14 17:20:02 ubuntu1 kernel: [29601.379193] bond0: link status definitely down for interface enp3s0, disabling it
Feb 14 17:20:02 ubuntu1 kernel: [29601.379199] bond0: first active interface up!
Feb 14 17:20:04 ubuntu1 kernel: [29603.064712] e1000e: enp3s0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: Rx/Tx
Feb 14 17:20:04 ubuntu1 kernel: [29603.079162] bond0: link status definitely up for interface enp3s0, 1000 Mbps full duplex


Is it a known problem? Apparently after some seconds the failed interface works well again. The problem doesn't occur very often.



In the file /proc/net/bonding/bond0 I can see, that the mode 4 was correctly recognized:



cat /proc/net/bonding/bond0 
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)

Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2 (0)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

802.3ad info
LACP rate: fast
Min links: 0
Aggregator selection policy (ad_select): stable
System priority: 65535


I have tried to use bond-slaves with interface names instead of none. But in this case the ifenslave was locked during restarting of network services. So I have found a recommendation, that using "none" the bond0 will be up and doesn't lock.



Any ideas?







networking network-bonding






share|improve this question















share|improve this question













share|improve this question




share|improve this question








edited Feb 14 at 17:18

























asked Feb 14 at 16:44









Andrej

64




64











  • Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
    – Terrance
    Feb 14 at 17:00










  • Do you have any suggestions how to trace or diagnostic this problem?
    – Andrej
    Feb 14 at 17:20










  • Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
    – Terrance
    Feb 14 at 17:22










  • This is Linksys LGS326 26-Port Gigabit Smart Switch.
    – Andrej
    Feb 14 at 17:28










  • I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
    – Terrance
    Feb 14 at 18:05
















  • Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
    – Terrance
    Feb 14 at 17:00










  • Do you have any suggestions how to trace or diagnostic this problem?
    – Andrej
    Feb 14 at 17:20










  • Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
    – Terrance
    Feb 14 at 17:22










  • This is Linksys LGS326 26-Port Gigabit Smart Switch.
    – Andrej
    Feb 14 at 17:28










  • I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
    – Terrance
    Feb 14 at 18:05















Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
– Terrance
Feb 14 at 17:00




Look at help.ubuntu.com/community/UbuntuBonding and maybe try changing the bond-mode to bond-mode 4 and add both NICs to the bond-slaves line. Did you also add bonding to the /etc/modules file? These are only suggestions as unless we can duplicate your environment some answers can only be guesses.
– Terrance
Feb 14 at 17:00












Do you have any suggestions how to trace or diagnostic this problem?
– Andrej
Feb 14 at 17:20




Do you have any suggestions how to trace or diagnostic this problem?
– Andrej
Feb 14 at 17:20












Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
– Terrance
Feb 14 at 17:22




Hmmmm, I am not 100% sure. Most of the systems I have had to deal with doing Link-aggregation were on Red Hat servers. However, I have seen that some of the Intel cards that use the e1000e you might have to get the network driver directly from Intel. I have also seen switches be problems for us, but not very often. What type of switch are you connecting to?
– Terrance
Feb 14 at 17:22












This is Linksys LGS326 26-Port Gigabit Smart Switch.
– Andrej
Feb 14 at 17:28




This is Linksys LGS326 26-Port Gigabit Smart Switch.
– Andrej
Feb 14 at 17:28












I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
– Terrance
Feb 14 at 18:05




I have always been a huge fan of Linksys products. The switch looks nice and has good setup pages on it. Have you tried maybe setting the LACP on the ports of the switch to Dynamic or Static?
– Terrance
Feb 14 at 18:05















active

oldest

votes











Your Answer







StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "89"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
convertImagesToLinks: true,
noModals: false,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













 

draft saved


draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2faskubuntu.com%2fquestions%2f1006197%2fproblems-with-bonding-on-16-04-lts%23new-answer', 'question_page');

);

Post as a guest



































active

oldest

votes













active

oldest

votes









active

oldest

votes






active

oldest

votes















 

draft saved


draft discarded















































 


draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2faskubuntu.com%2fquestions%2f1006197%2fproblems-with-bonding-on-16-04-lts%23new-answer', 'question_page');

);

Post as a guest













































































Popular posts from this blog

pylint3 and pip3 broken

Missing snmpget and snmpwalk

How to enroll fingerprints to Ubuntu 17.10 with VFS491