Running connectors on IBM z/OS

You can use the IBM MQ connectors to connect into IBM MQ for z/OS, and you can run the connectors on z/OS as well, connecting into the queue manager using bindings mode. These instructions explain how to run Kafka Connect in both standalone and distributed mode. For more information and to help decide which mode to use see the explanation of Kafka Connect workers.

Before you can run IBM MQ connectors on IBM z/OS, you must prepare your Kafka files and your system as follows.

Setting up Kafka to run on IBM z/OS

You can run Kafka Connect workers on IBM z/OS Unix System Services. To do so, you must ensure that the Kafka Connect shell scripts and the Kafka Connect configuration files are converted to EBCDIC encoding.

Download the Kafka Connect files

Download Apache Kafka to a non-z/OS system to retrieve the .tar file that includes the Kafka Connect shell scripts and JAR files.

To download Kafka Connect and make it available to your z/OS system:

  1. Log in to a system that is not running IBM z/OS, for example, a Linux system.
  2. Download Apache Kafka 3.4.0 or later to the system. Event Streams provides support for Kafka Connect if you are using a Kafka version listed in the Kafka version shipped column of the support matrix.
  3. Extract the downloaded .tgz file, for example:

    gunzip -k kafka_2.13-3.4.0.tgz
    
  4. Copy the resulting .tar file to a directory on the z/OS Unix System Services.

Download IBM MQ connectors and configuration

Depending on the connector you want to use:

If you want to run a standalone Kafka Connect worker, you need a .properties file. To run a distributed Kafka Connect worker, you need a .json file.

Copy the connector JAR file(s) and the required configuration file to a directory on the z/OS Unix System Services.

Convert the files

If you want to run a standalone Kafka Connect worker, convert the following shell scripts and configuration files from ISO8859-1 to EBCDIC encoding:

  • bin/kafka-run-class.sh
  • bin/connect-standalone.sh
  • config/connect-standalone.properties
  • mq-source.properties or mq-sink.properties

If you want to run a distributed Kafka Connect worker, convert the following shell scripts and configuration files from ISO8859-1 to EBCDIC encoding:

  • bin/kafka-run-class.sh
  • bin/connect-distributed.sh
  • config/connect-distributed.sh

Extract the Apache Kafka distribution:

  1. Log in to the IBM z/OS system and access the Unix System Services.
  2. Change to an empty directory that you want to use for the Apache Kafka distribution, and copy the .tar file to the new directory.
  3. Extract the .tar file, for example:

    tar -xvf kafka_2.13-2.8.1.tar
    
  4. Change to the resulting kafka_<version> directory.

Convert the shell scripts:

  1. Copy the connect-standalone.sh shell script (or connect-distributed.sh for a distributed setup) into the current directory, for example:

    cp bin/connect-standalone.sh ./connect-standalone.sh.orig
    
  2. Determine the codeset on the IBM z/OS system by running:

    locale -k codeset
    
  3. Convert the script to EBCDIC encoding and replace the original, for example for codeset IBM-1047:

    iconv -f ISO8859-1 -t IBM-1047 ./connect-standalone.sh.orig > bin/connect-standalone.sh
    
  4. Ensure the file permissions are set so that the script is executable, for example:

    chmod +x bin/connect-standalone.sh
    
  5. Copy the kafka-run-class.sh shell script into the current directory, for example:

    cp bin/kafka-run-class.sh ./kafka-run-class.sh.orig
    
  6. Convert the script to EBCDIC encoding and replace the original, for example for codeset IBM-1047:

    iconv -f ISO8859-1 -t IBM-1047 ./kafka-run-class.sh.orig > bin/kafka-run-class.sh
    
  7. Ensure the file permissions are set so that the script is executable, for example:

    chmod +x bin/kafka-run-class.sh
    

Convert the configuration files:

  1. Copy the connect-standalone.properties file (or connect-distributed.properties for a distributed setup) into the current directory, for example:

    cp config/connect-standalone.properties ./connect-standalone.properties.orig
    
  2. Determine the codeset on the IBM z/OS system by running:

    locale -k codeset
    
  3. Convert the script to EBCDIC encoding and replace the original, for example for codeset IBM-1047:

    iconv -f ISO8859-1 -t IBM-1047 ./connect-standalone.properties.orig > config/connect-standalone.properties
    

If running in standalone mode:

  1. Copy the MQ .properties file into the current directory, for example:

    cp ./mq-source.properties ./mq-source.properties.orig
    
  2. Determine the codeset on the IBM z/OS system by running:

    locale -k codeset
    
  3. Convert the script to EBCDIC encoding and replace the original, for example for codeset IBM-1047:

    iconv -f ISO8859-1 -t IBM-1047 ./mq-source.properties.orig > ./mq-source.properties
    

Note: For distributed mode the .json file must remain in ASCII format.

Update the Kafka Connect configuration

The connect-standalone.properties (or connect-distributed.properties for distributed mode) file must include the correct bootstrap.servers and SASL/SSL configuration for your Apache Kafka or Event Streams install.

For example, if running against Event Streams, download the certificate for your install to your IBM z/OS system. Generate credentials that can produce, consume and create topics and update the connect-standalone.properties (or connect-distributed.properties) file to include:

bootstrap.servers=<bootstrapServers>
security.protocol=SASL_SSL
ssl.protocol=TLSv1.2
ssl.truststore.location=/opt/kafka/es-cert.p12
ssl.truststore.password=<truststorePassword>
ssl.truststore.type=PKCS12
sasl.mechanism=SCRAM-SHA-512
sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="<userName>" password="<password>";
producer.security.protocol=SASL_SSL
producer.ssl.protocol=TLSv1.2
producer.ssl.truststore.location=/opt/kafka/es-cert.p12
producer.ssl.truststore.password=<truststorePassword>
producer.ssl.truststore.type=PKCS12
producer.sasl.mechanism=SCRAM-SHA-512
producer.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="userName" password="<password>";
consumer.security.protocol=SASL_SSL
consumer.ssl.protocol=TLSv1.2
consumer.ssl.truststore.location=/opt/kafka/es-cert.p12
consumer.ssl.truststore.password=<truststorePassword>
consumer.ssl.truststore.type=PKCS12
consumer.sasl.mechanism=SCRAM-SHA-512
consumer.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="userName" password="<password>";
plugin.path=/opt/connectors

Configuring the environment

The IBM MQ connectors use the JMS API to connect to MQ. You must set the environment variables required for JMS applications before running the connectors on IBM z/OS.

Ensure you set CLASSPATH to include com.ibm.mq.allclient.jar, and also set the JAR file for the connector you are using - this is the connector JAR file you downloaded from the Event Streams UI or built after cloning the GitHub project, for example, kafka-connect-mq-source-1.3.0-jar-with-dependencies.jar.

As you are using the bindings connection mode for the connector to connect to the queue manager, also set the following environment variables:

  1. The STEPLIB used at run time must contain the IBM MQ SCSQAUTH and SCSQANLE libraries. Specify this library in the startup JCL, or specify it by using the .profile file.

    From UNIX and Linux System Services, you can add these using a line in your .profile file as shown in the following code snippet, replacing thlqual with the high-level data set qualifier that you chose when installing IBM MQ:

    export STEPLIB=thlqual.SCSQAUTH:thlqual.SCSQANLE:$STEPLIB
    
  2. The connector needs to load a native library. Set LIBPATH to include the following directory of your MQ installation:

    <path_to_MQ_installation>/mqm/<MQ_version>/java/lib
    

The bindings connection mode is a configuration option for the connector as described in the source connector GitHub README and in the sink connector GitHub README.

Starting Kafka Connect on z/OS

Kafka Connect is started by using a bash script. If you do not already have bash installed on your z/OS system, install it now.

To install bash version 4.2.53 or later:

  1. Download the bash archive file from IBM Open Enterprise Foundation for z/OS..

  2. Extract the archive file to get the .tar file:

     gzip -d bash.tar.gz
    
  3. FTP the .tar file to your z/OS USS directory such as /bin
  4. Extract the .tar file to install bash:

    tar -cvfo bash.tar
    

If bash on your z/OS system is not in /bin, you need to update the kafka-run-class.sh file. For example, if bash is located in /usr/local/bin update the first line of kafka-run-class.sh to have #!/usr/local/bin/bash

Starting Kafka Connect in standalone mode

To start Kafka Connect in standalone mode navigate to your Kafka directory and run the connect-standalone.sh script, passing in your connect-standalone.properties and mq-source.properties or mq-sink.properties. For example:

cd kafka
./bin/connect-standalone.sh connect-standalone.properties mq-source.properties

For more details on creating the properties files see the connecting MQ documentation. Make sure connection type is set to bindings mode.

Starting Kafka Connect in distributed mode

To start Kafka Connect in distributed mode navigate to your Kafka directory and run the connect-distributed.sh script, passing in your connect-distributed.properties. Unlike in standalone mode, MQ properties are not passed in on startup. For example:

cd kafka
./bin/connect-distributed.sh connect-distributed.properties

To start an individual connector use the Kafka Connect REST API. For example, given a configuration file mq-source.json with the following contents:

{
    "name":"mq-source",
        "config" : {
            "connector.class":"com.ibm.eventstreams.connect.mqsource.MQSourceConnector",
            "tasks.max":"1",
            "mq.queue.manager":"QM1",
            "mq.connection.mode":"bindings",
            "mq.queue":"MYQSOURCE",
            "mq.record.builder":"com.ibm.eventstreams.connect.mqsource.builders.DefaultRecordBuilder",
            "topic":"test",
            "key.converter":"org.apache.kafka.connect.storage.StringConverter",
            "value.converter":"org.apache.kafka.connect.converters.ByteArrayConverter"
        }
    }

start the connector using:

curl -X POST http://localhost:8083/connectors -H "Content-Type: application/json" -d @mq-source.json

Advanced configuration

For more details about the connectors and to see all configuration options, see the source connector GitHub README or sink connector GitHub README.