Configuring Load Balancing as a Service in TF



Overview: Load Balancing as a Service

Load Balancing as a Service (LBaaS) is a feature available through OpenStack Neutron. Tungsten Fabric 1.20 and greater allows the use of the Neutron API for LBaaS to apply open source load balancing technologies to provision a load balancer in the TF system.

The LBaaS load balancer enables the creation of a pool of virtual machines serving applications, all front-ended by a virtual-ip. The LBaaS implementation has the following features:

  • Load balancing of traffic from clients to a pool of backend servers. The load balancer proxies all connections to its virtual IP.

  • Provides load balancing for HTTP, TCP, and HTTPS.

  • Provides health monitoring capabilities for applications, including HTTP, TCP, and ping.

  • Enables floating IP association to virtual-ip for public access to the backend pool.

In Figure 1, the load balancer is launched with the virtual IP address The backend pool of virtual machine applications (App Pool) is on the subnet Each of the application virtual machines gets an IP address (virtual-ip) from the pool subnet. When a client connects to the virtual-ip for accessing the application, the load balancer proxies the TCP connection on its virtual-ip, then creates a new TCP connection to one of the virtual machines in the pool.

The pool member is selected using one of following methods:

  • weighted round robin (WRR), based on the weight assignment

  • least connection, selects the member with the fewest connections

  • source IP selects based on the source-ip of the packet

Figure 1: Load Balancing as a Service in TF

Additionally, the load balancer monitors the health of each pool member using the following methods:

  • Monitors TCP by creating a TCP connection at intervals.

  • Monitors HTTP by creating a TCP connection and issuing an HTTP request at intervals.

  • Monitors ping by checking if a member can be reached by pinging.

TF LBaaS Implementation

TF supports the OpenStack LBaaS Neutron APIs and creates relevant objects for LBaaS, including virtual-ip, loadbalancer-pool, loadbalancer-member, and loadbalancer-healthmonitor. TF creates a service instance when a loadbalancer-pool is associated with a virtual-ip object. The service scheduler then launches a namespace on a randomly selected virtual router and spawns HAProxy into that namespace. The configuration for HAProxy is picked up from the load balancer objects. TF supports high availability of namespaces and HAProxy by spawning active and standby on two different vrouters.

A Note on Installation

To use the LBaaS feature, HAProxy, version 1.5 or greater and iproute2, version 3.10.0 or greater must both be installed on the TF compute nodes.

If you are using fab commands for installation, the haproxy and iproute2 packages will be installed automatically with LBaaS if you set the following:


Use the following to check the version of the iproute2 package on your system:

root@nodeh5:/var/log# ip -V
ip utility, iproute2-ss130716


LBaaS currently has these limitations:

  • A pool should not be deleted before deleting the VIP.

  • Multiple VIPs cannot be associated with the same pool. If pool needs to be reused, create another pool with the same members and bind it to the second VIP.

  • Members cannot be moved from one pool to another. If needed, first delete the members from one pool, then add to a different pool.

  • In case of active-standby failover, namespaces might not get cleaned up when the agent restarts.

  • The floating-ip association needs to select the VIP port and not the service ports. ​

Configuring LBaaS Using CLI

The LBaaS feature is enabled on TF through Neutron API calls. The following procedure shows how to create a pool network and a VIP network using CLI. The VIP network is created in the public network and members are added in the pool network.

Creating a Load Balancer

Use the following steps to create a load balancer in TF.

  1. Create a VIP network.

    neutron net-create vipnet

    neutron subnet-create –-name vipsubnet vipnet

  2. Create a pool network.

    neutron net-create poolnet

    neutron subnet-create --name poolsubnet poolnet

  3. Create a pool for HTTP.

    neutron lb-pool-create --lb-method ROUND_ROBIN --name mypool --protocol HTTP --subnet-id poolsubnet

  4. Add members to the pool.

    neutron lb-member-create --address --protocol-port 80 mypool

    neutron lb-member-create --address --protocol-port 80 mypool

  5. Create a VIP for HTTP and associate it to the pool.

    neutron lb-vip-create --name myvip --protocol-port 80 --protocol HTTP--subnet-id vipsubnet mypool

Deleting a Load Balancer

Use the following steps to delete a load balancer in TF.

  1. Delete the VIP.

    neutron lb-vip-delete <vip-uuid>

  2. Delete members from the pool.

    neutron lb-member-delete <member-uuid>

  3. Delete the pool.

    neutron lb-pool-delete <pool-uuid>

Managing Healthmonitor for Load Balancer

Use the following commands to create a healthmonitor, associate a healthmonitor to a pool, disassociate a healthmonitor, and delete a healthmonitor.

  1. Create a healthmonitor.

    neutron lb-healthmonitor-create --delay 20 --timeout 10 --max-retries 3 --type HTTP

  2. Associate a healthmonitor to a pool.

    neutron lb-healthmonitor-associate <healthmonitor-uuid> mypool

  3. Disassociate a healthmonitor from a pool.

    neutron lb-healthmonitor-disassociate <healthmonitor-uuid> mypool

Configuring an SSL VIP with an HTTP Backend Pool

Use the following steps to configure an SSL VIP with an HTTP backend pool.

  1. Copy an SSL certificate to all compute nodes.

    scp ssl_certificate.pem <compute-node-ip> <certificate-path>

  2. Update the information in/etc/contrail/contrail-vrouter-agent.conf.

    # SSL certificate path haproxy


  3. Restart contrail-vrouter-agent.

    service contrail-vrouter-agent restart

  4. Create a VIP for port 443 (SSL)​.

    neutron lb-vip-create --name myvip --protocol-port 443 --protocol HTTP --subnet-id vipsubnet mypool​