Adding worker nodes to Classic clusters
Classic infrastructure
Reviewing the following sections for information on how to add worker nodes to your Classic cluster.
Looking to add workers to VPC clusters? See Adding worker nodes to VPC clusters.
To increase the availability of your apps, you can add worker nodes to an existing zone or multiple existing zones in your cluster. To help protect your apps from zone failures, you can add zones to your cluster.
When you create a cluster, the worker nodes are provisioned in a worker pool. After cluster creation, you can add more worker nodes to a pool by resizing it or by adding more worker pools. By default, the worker pool exists in one zone. Clusters that have a worker pool in only one zone are called single zone clusters. When you add more zones to the cluster, the worker pool exists across the zones. Clusters that have a worker pool that is spread across more than one zone are called multizone clusters.
If you have a multizone cluster, keep its worker node resources balanced. Make sure that all the worker pools are spread across the same zones, and add or remove workers by resizing the pools instead of adding individual nodes. After you set up your worker pool, you can set up the cluster autoscaler to automatically add or remove worker nodes from your worker pools based on your workload resource requests.
Want to save on your classic worker node costs? Create a reservation to lock in a discount over 1 or 3 year terms! Then, create your worker pool by using the reserved instances. Note that autoscaling can't be enabled on worker pools that use reservations.
Creating a new worker pool
You can add worker nodes to your classic cluster by creating a new worker pool.
Before you begin, make sure that you have the [Operator or Administrator IAM platform access role](/docs/openshift?topic=openshift-iam-platform-access-roles.
-
Retrieve the Worker Zones of your cluster and choose the zone where you want to deploy the worker nodes in your worker pool. If you have a single zone cluster, you must use the zone that you see in the Worker Zones field. For multizone clusters, you can choose any of the existing Worker Zones of your cluster, or add one of the multizone locations for the region that your cluster is in. You can list available zones by running
ibmcloud oc zone ls
.ibmcloud oc cluster get --cluster <cluster_name_or_ID>
Example output
... Worker Zones: dal10, dal12, dal13
-
For each zone, list available private and public VLANs. Note the private and the public VLAN that you want to use. If you don't have a private or a public VLAN, the VLAN is automatically created for you when you add a zone to your worker pool.
ibmcloud oc vlan ls --zone <zone>
-
For each zone, review the available flavors for worker nodes.
ibmcloud oc flavors --zone <zone>
-
Create a worker pool. For more options, see the CLI documentation.
- Include the
--label
option to automatically label worker nodes that are in the pool with the labelkey=value
. - If you provision a bare metal or dedicated VM worker pool, specify
--hardware dedicated
. - The new worker nodes run the same
major.minor
version as the cluster master, but the latest worker node patch of thatmajor.minor
version.
ibmcloud oc worker-pool create classic --name <pool_name> --cluster <cluster_name_or_ID> --flavor <flavor> --size-per-zone <number_of_workers_per_zone> [--operating-system (REDHAT_8_64)] [--label key=value]
- Include the
-
Verify that the worker pool is created.
ibmcloud oc worker-pool ls --cluster <cluster_name_or_ID>
-
By default, adding a worker pool creates a pool with no zones. To deploy worker nodes in a zone, you must add the zones that you previously retrieved to the worker pool. If you want to spread your worker nodes across multiple zones, repeat this command for each zone.
ibmcloud oc zone add classic --zone <zone> --cluster <cluster_name_or_ID> --worker-pool <pool_name> --private-vlan <private_VLAN_ID> --public-vlan <public_VLAN_ID>
-
Verify that worker nodes provision in the zone that you added. Your worker nodes are ready when the status changes from
provision_pending
tonormal
.ibmcloud oc worker ls --cluster <cluster_name_or_ID> --worker-pool <pool_name>
Example output
ID Public IP Private IP Machine Type State Status Zone Version kube-dal10-crb20b637238ea471f8d4a8b881aae4962-w7 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 provision_pending Ready dal10 1.31 kube-dal10-crb20b637238ea471f8d4a8b881aae4962-w8 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 provision_pending Ready dal10 1.31
Resizing a worker pool
You can add or reduce the number of worker nodes in your cluster by resizing an existing worker pool, regardless of whether the worker pool is in one zone or spread across multiple zones.
For example, consider a cluster with one worker pool that has three worker nodes per zone.
- If the cluster is single zone and exists in
dal10
, then the worker pool has three worker nodes indal10
. The cluster has a total of three worker nodes. - If the cluster is multizone and exists in
dal10
anddal12
, then the worker pool has three worker nodes indal10
and three worker nodes indal12
. The cluster has a total of six worker nodes.
For bare metal worker pools, keep in mind that billing is monthly. If you resize up or down, it impacts your costs for the month. When you add worker nodes by resizing a worker pool, the new worker nodes run the same major.minor
version as the cluster master, but the latest worker node patch of that major.minor
version.
Before you begin, make sure that you have the [Operator or Administrator IAM platform access role](/docs/openshift?topic=openshift-iam-platform-access-roles.
To resize the worker pool, change the number of worker nodes that the worker pool deploys in each zone:
- Get the name of the worker pool that you want to resize.
ibmcloud oc worker-pool ls --cluster <cluster_name_or_ID>
- Resize the worker pool by designating the number of worker nodes that you want to deploy in each zone.
ibmcloud oc worker-pool resize --cluster <cluster_name_or_ID> --worker-pool <pool_name> --size-per-zone <number_of_workers_per_zone>
- Verify that the worker pool is resized.
Example output for a worker pool that is in two zones,ibmcloud oc worker ls --cluster <cluster_name_or_ID> --worker-pool <pool_name>
dal10
anddal12
, and is resized to two worker nodes per zone:ID Public IP Private IP Machine Type State Status Zone Version kube-dal10-crb20b637238ea471f8d4a8b881aae4962-w7 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 normal Ready dal10 1.31 kube-dal10-crb20b637238ea471f8d4a8b881aae4962-w8 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 normal Ready dal10 1.31 kube-dal12-crb20b637238ea471f8d4a8b881aae4962-w9 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 normal Ready dal12 1.31 kube-dal12-crb20b637238ea471f8d4a8b881aae4962-w10 169.xx.xxx.xxx 10.xxx.xx.xxx b3c.4x16 normal Ready dal12 1.31
Adding a zone to a worker pool
You can span your classic cluster across multiple zones within one region by adding a zone to your existing worker pool.
When you add a zone to a worker pool, the worker nodes that are defined in your worker pool are provisioned in the new zone and considered for future workload scheduling. Red Hat OpenShift on IBM Cloud automatically adds the failure-domain.beta.kubernetes.io/region
label for the region and the failure-domain.beta.kubernetes.io/zone
label for the zone to each worker node. The Kubernetes scheduler uses these labels to spread pods across zones within the same region.
If you have multiple worker pools in your cluster, add the zone to all them so that worker nodes are spread evenly across your cluster. Note that when you add worker nodes to your cluster, the new worker nodes run the same major.minor
version as the cluster master, but the latest worker node patch of that major.minor
version.
Before you begin:
- To add a zone to your worker pool, your worker pool must be in a multizone-capable zone. If your worker pool is not in a multizone-capable zone, consider creating a new worker pool.
- Make sure that you have the [Operator or Administrator IAM platform access role](/docs/openshift?topic=openshift-iam-platform-access-roles.
- In classic clusters, if you have multiple VLANs for your cluster, multiple subnets on the same VLAN, or a multizone classic cluster, you must enable a Virtual Router Function (VRF) for your IBM Cloud infrastructure account so your worker nodes can communicate with each other on the private network. To enable VRF, see Enabling VRF. To check
whether a VRF is already enabled, use the
ibmcloud account show
command. If you can't or don't want to enable VRF, enable VLAN spanning. To perform this action, you need the Network > Manage Network VLAN Spanning infrastructure permission, or you can request the account owner to enable it. To check whether VLAN spanning is already enabled, use theibmcloud oc vlan spanning get --region <region>
command.
To add a zone with worker nodes to your worker pool:
-
List available zones and pick the zone that you want to add to your worker pool. The zone that you choose must be a multizone-capable zone.
ibmcloud oc zone ls
-
List available VLANs in that zone. If you don't have a private or a public VLAN, the VLAN is automatically created for you when you add a zone to your worker pool.
ibmcloud oc vlan ls --zone <zone>
-
List the worker pools in your cluster and note their names.
ibmcloud oc worker-pool ls --cluster <cluster_name_or_ID>
-
Add the zone to your worker pool. If you have multiple worker pools, add the zone to all your worker pools so that your cluster is balanced in all zones.
A private and a public VLAN must exist before you can add a zone to multiple worker pools. If you don't have a private and a public VLAN in that zone, add the zone to one worker pool first so that a private and a public VLAN is created for you. Then, you can add the zone to other worker pools by specifying the private and the public VLAN that was created for you.
If you want to use different VLANs for different worker pools, repeat this command for each VLAN and its corresponding worker pools. Any new worker nodes are added to the VLANs that you specify, but the VLANs for any existing worker nodes are not changed.
ibmcloud oc zone add classic --zone <zone> --cluster <cluster_name_or_ID> -p <pool_name> [-p <pool2_name>] --private-vlan <private_VLAN_ID> --public-vlan <public_VLAN_ID>
-
Verify that the zone is added to your cluster. Look for the added zone in the Worker zones field of the output. Note that the total number of workers in the Workers field has increased as new worker nodes are provisioned in the added zone.
ibmcloud oc cluster get --cluster <cluster_name_or_ID>
Example output
NAME: mycluster ID: df253b6025d64944ab99ed63bb4567b6 State: normal Status: healthy cluster Created: 2018-09-28T15:43:15+0000 Location: dal10 Pod Subnet: 172.30.0.0/16 Service Subnet: 172.21.0.0/16 Master URL: https://c3.<region>.containers.cloud.ibm.com:30426 Public Service Endpoint URL: https://c3.<region>.containers.cloud.ibm.com:30426 Private Service Endpoint URL: https://c3-private.<region>.containers.cloud.ibm.com:31140 Master Location: Dallas Master Status: Ready (21 hours ago) Ingress Subdomain: mycluster-<hash>-0000.us-south.containers.appdomain.cloud Ingress Secret: mycluster-<hash>-0000 Workers: 6 Worker Zones: dal10, dal12 Version: 1.31_1524 Owner: owner@email.com Resource Group ID: a8a12accd63b437bbd6d58fb6a462ca7 Resource Group Name: Default