To enable multiple availability zones for IBM Event Streams, ensure that your Kubernetes cluster is zone aware and that Kafka rack awareness is set up correctly.
For guidance about handling outages in a multizone setup, see managing a multizone setup.
Kubernetes uses zone-aware information to determine the zone location of each of its nodes in the cluster to enable scheduling of pod replicas in different zones.
Some clusters, typically AWS, will already be zone aware. For clusters that are not zone aware, each Kubernetes node will need to be set up with a zone label.
To determine if your cluster is zone aware:
- Log in to your Red Hat OpenShift Container Platform as a cluster administrator by using the
Run the following command as cluster administrator:
oc get nodes --show-labels
If your Kubernetes cluster is zone aware, the following label is displayed against each node:
topology.kubernetes.io/zoneif using OpenShift 4.5 or later
failure-domain.beta.kubernetes.io/zoneif using an earlier version of OpenShift
The value of the label is the zone the node is in, for example,
If your Kubernetes cluster is not zone aware, all cluster nodes will need to be labeled using a value that identifies the zone that each node is in. For example, run the following command to label and allocate a node to
oc label node <node-name> topology.kubernetes.io/zone=es-zone-1
The zone label is needed to set up rack awareness when installing for multizone.
Kafka rack awareness
In addition to zone awareness, Kafka rack awareness helps to spread the Kafka broker pods and Kafka topic replicas across different availability zones, and also sets the brokers’
broker.rack configuration property for each Kafka broker.
To set up Kafka rack awareness, Kafka brokers require a cluster role to provide permission to view which Kubernetes node they are running on.
Before applying Kafka rack awareness to an Event Streams installation, apply a cluster role and a cluster role binding:
Create a file called
eventstreams-kafka-broker.yamland copy the following YAML content to create the cluster role:
kind: ClusterRole apiVersion: rbac.authorization.k8s.io/v1 metadata: name: eventstreams-kafka-broker labels: app: eventstreams rules: - verbs: - get - create - watch - update - delete - list apiGroups: - rbac.authorization.k8s.io resources: - clusterrolebindings - verbs: - get - create - watch - update - delete - list apiGroups: - "" resources: - nodes
Apply the cluster role by using the following command:
oc apply -f eventstreams-kafka-broker.yaml
Create a file called
eventstreams-kafka-broker-crb.yamland copy the following YAML content to create the cluster role binding:
kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1 metadata: name: eventstreams-kafka-broker subjects: - kind: ServiceAccount name: eventstreams-cluster-operator namespace: <operator_namespace> roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: eventstreams-kafka-broker
Apply the cluster role binding by using the following command:
oc apply -f eventstreams-kafka-broker-crb.yaml