Configure Network Bonding on CentOS 6 / RHEL 6

Network Bonding is the aggregation of multiple lan cards(eth0 & eth1) into a single interface called bond interface. Network Bonding is a kernel feature and also known as NIC teaming. Bonding is generally required to provide high availability and load balancing services.

In this tutorial we will configure network bonding on CentOS 6.X and RHEL 6.X , in my case i have two Lan Cards (eth0 & eth1) & will create bond interface (bond0).

Step:1 Create the bond file ( ifcfg-bond0 ) and specify the ip address,netmask & gateway

# vi /etc/sysconfig/network-scripts/ifcfg-bond0


Step:2 Edit the files of eth0 & eth1 and make sure you enter the master and slave entry, as shown below

# vi /etc/sysconfig/network-scripts/ifcfg-eth0


# vi /etc/sysconfig/network-scripts/ifcfg-eth1


Step:3 Create the Bond file(bonding.conf)

# vi /etc/modprobe.d/bonding.conf

alias bond0 bonding
options bond0 mode=1 miimon=100

Different Modes used in bonding.conf file .

  • balance-rr or 0 — round-robin mode for fault tolerance and load balancing.
  • active-backup or 1 — Sets active-backup mode for fault tolerance.
  • balance-xor or 2 — Sets an XOR (exclusive-or) mode for fault tolerance and load balancing.
  • broadcast or 3 — Sets a broadcast mode for fault tolerance. All transmissions are sent on all slave interfaces.
  • 802.3ad or 4 — Sets an IEEE 802.3ad dynamic link aggregation mode. Creates aggregation groups that share the same speed & duplex settings.
  • balance-tlb or 5 — Sets a Transmit Load Balancing (TLB) mode for fault tolerance & load balancing.
  • balance-alb or 6 — Sets an Active Load Balancing (ALB) mode for fault tolerance & load balancing.

Step:4 Now Restart the network Service

# service network restart

Using below command to  Check the bond Interface

root@localhost ~]# ifconfig bond0
bond0     Link encap:Ethernet  HWaddr 08:00:27:5C:A8:8F
inet addr:  Bcast:  Mask:
inet6 addr: fe80::a00:27ff:fe5c:a88f/64 Scope:Link
RX packets:6164 errors:0 dropped:0 overruns:0 frame:0
TX packets:1455 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:482336 (471.0 KiB)  TX bytes:271221 (264.8 KiB

Step:5 Verify the Status of bond interface.

[root@localhost ~]# cat /proc/net/bonding/bond0  
Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)

Bonding Mode: fault-tolerance (active-backup)
Primary Slave: None
Currently Active Slave: eth0
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

Slave Interface: eth0
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 08:00:27:5c:a8:8f
Slave queue ID: 0

Slave Interface: eth1
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 08:00:27:7f:04:49
Slave queue ID: 0

Note: As per the output shown above , we are currently using active-backup bonding. To do the testing just disable one interface and check whether you are still accessing your machine using the bond ip.

6 Responses

  1. Navneet says:

    When I did one of the nic(eth0) down, bonding stopped working but /proc/net/bonding/bond0 was showing everything fine and bond0 was showing IP too. As soon as I enabled eth0, server again started pinging. Please help

  2. Hi Navneet ,

    Please check whether second interface like eth1 is detected and its link is up using the command “ethtool eth1”.

  3. Chris Enslow says:

    I’m noticing in RHEL documentation that they are recommending *NOT* to use modeprobe.conf with later RHEL releases – they recommend using “BONDING_OPTS=” in the ifcfg-bond[#] file instead so that it’s all in one place. It is supposed to work in later kernel releases.

  4. kiran says:

    For round-robin if have some problem. if I down the the nic1 from the console there is no interruption in my putty session, but if I down nic2 then my putty session becoming in active until I restart the network service. Do we really need to restart the network service once the nic goes bad?

    • Hi Kiran,

      Are you testing bonding on Virtual Machines hosted on Virtual Box …?

      If you want to achieve fault tolerance and load balancing then either go for mode 5 or mode 6. As the name suggests in round robin mode packets will flow in round robin way.

  1. August 31, 2015

    […] Please refer this for Different Modes in NIC bonding […]

Leave a Reply

Your email address will not be published. Required fields are marked *