PlenusLB is a bare metal and cloud load balancer for Kubernetes clusters. It is designed for environments where a cloud load balancer already integrated for Kubernetes is not available.
The current version support two scenarios:
- pure bare metal deployment
- integration with Hetzner cloud
PlenusLB has been originally developed to be used on the Plenus cloud platform and in bare metal environments.
PlenusLB works by taking IP addresses, defined in a IPPool custom resource, and assigning them to kubernetes services of type LoadBalancer. The allocation is managed with a IPAllocation custom resource; once assigned the IP, the status of the service is updated, then Kubernetes will create the iptables/ipvs rules to route the ingress traffic to the service.
PlenusLB also takes care of:
- when the IP pool declares a cloud provider, requesting the IP address to that cloud provider and, on the provider side, direct the routing to the server where the IP will be assigned
- choosing a cluster node to act as ingress node
- assigning the IP address to a given network interface of the node, this way the node will accept traffic for the IP; usually the interface is an empty bridge.
One of the design choices made was to use a dedicated bridge interface for publishing load balancer IP addresses on the nodes. This approach was chosen because it is agnostic towards the cloud provider, bare metal or virtual environments and requires no integration with routers, for BGP, or other network equipment.
PlenusLB has two components:
- the controller, one replica per cluster
- the operators, one for each worker node of the Kubernetes cluster
The controller orchestrates all operations by watching all resources (IP pools, allocations, load balancer services) and listening for kubernetes events in order to deal with error situations such as a node crash. When a node failure occurs, all allocations on that node are moved to a healthy node. The operators are intended to assign and remove IP addresses to and from the network interface of the node on which they are running, as required by the controller.
If no pool has option
options:
hostNetworkInterface:
addAddressesToInterface: true
operators will not be deployed.
For the bare metal scenario it will be necessary to reserve a pool of IP addresses for each cluster/PlenusLB: the addresses must be declared in a PersistentIPPool.
To use PlenusLB with the Hetzner cloud provider you will need to have a project active on the cloud, create an API key in the "API TOKENS" sections of the interface and specify this token in the IP pools. The kubernetes cluster where PlenusLB is operating needs to be in the same Hetzner cloud project.
At the moment PlenusLB will implement load balancers using Hetzner Floating IPs, it will not use Hetzner Load Balancers.
All cluster nodes need to have an interface which can be used to assign IP addresses to.
Since on that interface PlenusLB will remove all IP addresses during operator startup on each node, do not use any interface where normal IP addresses have been assigned to the node.
We strongly suggest to use a bridge specifically created for this purpose, for example pl0
On Ubuntu/Debian nodes an empty (without physical interfaces) bridge pl0
can be created with the following commands:
apt-get update && \
apt-get -q -y install bridge-utils bash && \
bash -c \"echo -e 'auto pl0\niface pl0 inet manual\n bridge_ports none\n bridge_stp off\n bridge_fd 0\n bridge_maxwait 0' > /etc/network/interfaces.d/90-bridge-pl0.cfg\" && \
/etc/init.d/networking restart
The bridge will be persistent and will be restarted automatically after reboots, it is sufficient to give the commands once.
Using cloud-init put the following lines in cloud-config:
#cloud-config
bootcmd:
- [ cloud-init-per, once, plenuslb-init, sh, -xc, "apt-get update && apt-get -q -y install bridge-utils bash && bash -c \"echo -e 'auto pl0\niface pl0 inet manual\n bridge_ports none\n bridge_stp off\n bridge_fd 0\n bridge_maxwait 0' > /etc/network/interfaces.d/90-bridge-pl0.cfg\" && /etc/init.d/networking restart" ]
PlenusLB can be installed with the helm chart in the Plenus helm chart repository. Installation has been tested with helm 3.x and it is assumed a namespace plenuslb has been already created.
helm repo add plenus https://plenus-charts.storage.googleapis.com/stable/
helm repo update
helm upgrade --install \
--namespace=plenuslb \
--set envs.CLUSTER_NAME=mycluster \
--atomic --wait plenuslb plenus/plenuslb
The same command can be used to upgrade PlenusLB.
The value specified in the CLUSTER_NAME variable will be used as a part of the name given to the ephemeral IP addresses created on a cloud provider. Set it to a value related to the cluster where PlenusLB is installed, so that the IP addresses can be easily identified with the cluster.
Docker images are published in the following repositories:
- controller: https://hub.docker.com/repository/docker/plenus/plenuslb
- operator: https://hub.docker.com/repository/docker/plenus/plenuslb-operator
PlenusLB supports two types of IP pools:
- ephemeral IP pools, useful when the life cycle of the IP follows the life cycle of the service
- persistent IP pools, targeted for all those cases where a static reservation of the IP is necessary
At the moment it is not possible to create an IP address with an ephemeral IP pool and then migrate it to a persistent IP pool.
By creating a service with type: LoadBalancer and not specifying any externalIPs PlenusLB will provision an ephemeral IP: the IP will be assigned to the service as long as the service exists, but there is no reservation; if the service is deleted the IP will be released on the cloud provider. Ephemeral IP addresses cannot be used in the bare metal scenario.
To use ephemeral IP addresses it is necessary to create an EphemeralIPPool:
apiVersion: loadbalancing.plenus.io/v1alpha1
kind: EphemeralIPPool
metadata:
name: hetzner-eph-pool-all
spec:
cloudIntegration:
hetzner:
token: YOUR_HETZNER_API_TOKEN
options:
hostNetworkInterface:
addAddressesToInterface: true
interfaceName: pl0
cloudIntegration
declares the cloud provider where PlenusLB will create the IP addresses. At the moment only hetzner
is supported, and accepts a single parameter token
which must contain an Hetzner API key; the IP addresses will be created in the project that the API keys are authorized for, this must be the same project where the kubernetes cluster has been created.
options.hostNetworkInterface.interfaceName
must be set to the interface name where PlenusLB will assign IP addresses. Mandatory if addAddressesToInterface
is true.
options.hostNetworkInterface.addAddressesToInterface
usually is set to true, if set to false PlenusLB would not perform the assignment of the IP address to any interface on the ingress node; in that case the IP address would have to be assigned to the node manually or by another component.
To have an ephemeral IP assigned create a service with type: LoadBalancer and no externalIPs
apiVersion: v1
kind: Service
metadata:
name: hello-kubernetes
spec:
type: LoadBalancer
ports:
- port: 80
targetPort: 8080
selector:
app: hello-kubernetes
apiVersion: apps/v1
kind: Deployment
metadata:
name: hello-kubernetes
spec:
replicas: 1
selector:
matchLabels:
app: hello-kubernetes
template:
metadata:
labels:
app: hello-kubernetes
spec:
containers:
- name: hello-kubernetes
image: paulbouwer/hello-kubernetes:1.8
ports:
- containerPort: 8080
Persistent IP addresses can be used in all those cases where a reservation for the IP is desiderable, regardless where there is a service requesting the IP or not. Moreover at the moment it is the only type of IP supported in the bare metal scenario.
The idea is that the pool of available IP addresses is preassigned:
- in the bare metal scenario a pool of IP addresses is reserved for the kubernetes cluster
- on a cloud provider the IP addresses will be manually acquired, for example on Hetzner you will have to buy the IP addresses in the project accessible to PlenusLB through the given token
Then it is necessary to create a PersistentIPPool containing the addresses.
The following example is for Hetzner:
apiVersion: loadbalancing.plenus.io/v1alpha1
kind: PersistentIPPool
metadata:
name: hetzner-persist-pool-all
spec:
addresses:
- "1.2.3.4"
- "1.2.3.5"
cloudIntegration:
hetzner:
token: YOUR_HETZNER_API_TOKEN
options:
hostNetworkInterface:
addAddressesToInterface: true
interfaceName: pl0
For the bare metal case omit the cloudIntegration section.
Every persistent IP can be bound to a single service.
To have PlenusLB assign the persistent IP to a service it is sufficient to specify it as an externalIPs in a LoadBalancer type service.
apiVersion: v1
kind: Service
metadata:
name: hello-kubernetes-persistent
spec:
type: LoadBalancer
externalIPs:
- 1.2.3.4
ports:
- port: 80
targetPort: 8080
selector:
app: hello-kubernetes
Where 1.2.3.4 is the IP that you want assigned to the service.
PlenusLB provides some degrees of multi tenancy: if a cluster has multiple users, each one of them confined to a set of namespaces, it it possible to create IP pools reserved for specific namespaces. This, combined with the use of persistent IP pools, allows to allocate some IP addresses for specific users/projects.
The following PersistentIPPool declares two IP addresses that can be requested only from services in the namespaces project1 and project2. The PersistentIPPool does not declare any cloudIntegration, so it is for bare metal environment.
apiVersion: loadbalancing.plenus.io/v1alpha1
kind: PersistentIPPool
metadata:
name: baremetal-persist-pool-reserved
spec:
allowedNamespaces:
- project1
- project2
addresses:
- "1.2.3.6"
- "1.2.3.7"
options:
hostNetworkInterface:
addAddressesToInterface: true
interfaceName: pl0
The same notation can be used with EphemeralIPPool. It could be useful when there are multiple projects on the same cluster and only some projects must be allowed to request IP addresses from the cloud provider.
The controller deployment and the operator daemonset use a health check port; the default value for this port is 8080.
Since the operator use hostNetwork: true it needs this port on the nodes for exclusive use. In some rare cases there can be other pods using hostNetwork that need this port. For example Ceph CSI driver https://github.com/ceph/ceph-csi
The port is configurable with the HEALTH_PORT env variable in the controller. The controller will propagate it to the operator daemonset.
To build a specified tag:
git checkout v0.3.5
./build.sh -v v0.3.5
We welcome contributions in any form. If you want to contribute to the code base, for example to add another cloud provider, you can fork the project and then submit a pull request with your changes.