kafka¤
This role provides support to install a Kafka Cluster using Strimzi, Red Hat AMQ Streams, IBM Event Streams or AWS MSK and generate configuration that can be directly applied to Maximo Application Suite.
Both Strimzi and Red Hat AMQ streams component are massively scalable, distributed, and high-performance data streaming platform based on the Apache Kafka project. Both offer a distributed backbone that allows microservices and other applications to share data with high throughput and low latency.
As more applications move to Kubernetes and Red Hat OpenShift, it is increasingly important to be able to run the communication infrastructure on the same platform. Red Hat OpenShift, as a highly scalable platform, is a natural fit for messaging technologies such as Kafka. The AMQ streams component makes running and managing Apache Kafka OpenShift native through the use of powerful operators that simplify the deployment, configuration, management, and use of Apache Kafka on Red Hat OpenShift.
The AMQ streams component is part of the Red Hat AMQ family, which also includes the AMQ broker, a longtime innovation leader in Java™ Message Service (JMS) and polyglot messaging, as well as the AMQ interconnect router, a wide-area, peer-to-peer messaging solution. Under the covers, AMQ streams leverages Strimzi's architecture, resources and configurations.
Note: The MAS license does not include entitlement for AMQ streams. The MAS Devops Collection supports this Kafka deployment as an example only. Therefore, we recommend the use of Strimzi for an opensource Kafka provider.
Tip
The role will generate a yaml file containing the definition of a Secret and KafkaCfg resource that can be used to configure the deployed cluster as the MAS system Kafka.
This file can be directly applied using oc apply -f $MAS_CONFIG_DIR/kafkacfg-amqstreams-system.yaml or used in conjunction with the suite_config role.
Role Variables¤
General Variables¤
kafka_action¤
Specifies which operation to perform on the Kafka cluster.
- Optional
- Environment Variable:
KAFKA_ACTION - Default Value:
install
Purpose: Controls what action the role executes against Kafka deployments. This allows the same role to handle installation, upgrades, and removal of Kafka clusters.
When to use:
- Use install (default) for initial Kafka deployment
- Use upgrade to upgrade existing Kafka cluster (Strimzi and Red Hat AMQ Streams only)
- Use uninstall to remove Kafka cluster and operator
Valid values: install, upgrade, uninstall
Impact:
- install: Deploys Kafka operator and creates cluster
- upgrade: Upgrades existing Kafka cluster (only for Strimzi and Red Hat providers)
- uninstall: Removes Kafka cluster and operator
Related variables:
- kafka_provider: Upgrade action only supported for strimzi and redhat providers
- kafka_version: Target version when upgrading
Note: The upgrade action is only available for Strimzi and Red Hat AMQ Streams providers. IBM Event Streams and AWS MSK do not support in-place upgrades through this role.
kafka_provider¤
Specifies which Kafka provider to use for deployment.
- Optional
- Environment Variable:
KAFKA_PROVIDER - Default Value:
strimzi
Purpose: Determines which Kafka implementation to deploy. Different providers offer different features, licensing models, and deployment targets (on-cluster vs cloud-managed).
When to use:
- Use strimzi (default) for open-source Kafka on OpenShift (recommended, no additional license required)
- Use redhat for Red Hat AMQ Streams (requires separate license not included with MAS)
- Use ibm for IBM Event Streams on IBM Cloud (managed service, additional cost)
- Use aws for AWS MSK (managed service, additional cost)
Valid values: strimzi, redhat, ibm, aws
Impact:
- strimzi: Deploys open-source Kafka using Strimzi operator (no additional license)
- redhat: Deploys AMQ Streams (requires Red Hat AMQ license)
- ibm: Provisions IBM Event Streams in IBM Cloud (requires IBM Cloud account and incurs costs)
- aws: Provisions AWS MSK in AWS account (requires AWS account and incurs costs)
Related variables:
- Different providers require different additional variables
- kafka_action=upgrade only supported for strimzi and redhat
Note: IMPORTANT - MAS license does NOT include entitlement for Red Hat AMQ Streams. Strimzi is recommended for open-source Kafka. IBM and AWS providers provision managed cloud services with additional costs.
Red Hat AMQ Streams & Strimzi Variables¤
kafka_version¤
Kafka version to deploy (Strimzi and Red Hat AMQ Streams only).
- Optional
- Environment Variable:
KAFKA_VERSION - Default Value:
3.8.0for AMQ Streams,3.9.0for Strimzi
Purpose: Specifies which Apache Kafka version to deploy when using Strimzi or Red Hat AMQ Streams providers. The version must be supported by the installed operator.
When to use: - Leave as default for standard deployments - Set explicitly when you need a specific Kafka version - Verify version compatibility with operator before changing
Valid values: Valid Kafka version supported by the operator (e.g., 3.8.0, 3.9.0)
Impact: Determines which Kafka version is deployed. Incompatible versions will cause deployment failures.
Related variables:
- kafka_provider: Only applies to strimzi and redhat providers
- Operator version determines available Kafka versions
Note: Before changing this value, verify the version is supported by your AMQ Streams operator or Strimzi operator. This variable does not apply to IBM Event Streams or AWS MSK providers.
kafka_namespace¤
OpenShift namespace where Kafka operator and cluster will be deployed.
- Optional
- Environment Variable:
KAFKA_NAMESPACE - Default Value:
amq-streamsfor AMQ Streams,strimzifor Strimzi
Purpose: Specifies the namespace for deploying the Kafka operator and Kafka cluster resources. This isolates Kafka resources from other applications.
When to use: - Use default for standard single-cluster deployments - Set to custom namespace when organizing multiple Kafka deployments - Must be unique if deploying multiple Kafka clusters
Valid values: Any valid Kubernetes namespace name (e.g., strimzi, amq-streams, kafka-prod)
Impact: All Kafka resources (operator, cluster, topics, users) are created in this namespace.
Related variables:
- kafka_provider: Default namespace depends on provider (strimzi or amq-streams)
- kafka_cluster_name: Cluster created within this namespace
Note: The default namespace differs by provider: strimzi for Strimzi provider, amq-streams for Red Hat AMQ Streams. This variable only applies to Strimzi and Red Hat providers.
kafka_cluster_name¤
Name for the Kafka cluster resource.
- Optional
- Environment Variable:
KAFKA_CLUSTER_NAME - Default Value:
maskafka
Purpose: Defines the name of the Kafka custom resource that will be created. This name is used to identify the Kafka cluster and is embedded in resource names.
When to use:
- Use default (maskafka) for standard MAS deployments
- Set to custom name when deploying multiple Kafka clusters
- Use descriptive names for multi-cluster environments
Valid values: Valid Kubernetes resource name (lowercase alphanumeric with hyphens, e.g., maskafka, kafka-prod, mas-kafka)
Impact: This name is used throughout the Kafka deployment in resource names (services, pods, secrets). It appears in the generated KafkaCfg for MAS.
Related variables:
- kafka_namespace: Cluster created within this namespace
- Used in generated KafkaCfg when mas_instance_id is provided
Note: Choose a meaningful name as it appears in many resource names. The default maskafka is suitable for single-cluster MAS deployments.
kafka_cluster_size¤
Predefined configuration size for the Kafka cluster.
- Optional
- Environment Variable:
KAFKA_CLUSTER_SIZE - Default Value:
small
Purpose: Selects a predefined resource configuration for the Kafka cluster. Different sizes allocate different amounts of CPU, memory, and replicas for Kafka brokers and Zookeeper nodes.
When to use:
- Use small (default) for development, test, or small production environments
- Use large for production environments with higher throughput requirements
- Choose based on expected message volume and performance needs
Valid values: small, large
Impact:
- small: Fewer resources, suitable for dev/test or small workloads
- large: More resources (CPU, memory, replicas) for production workloads
Related variables:
- kafka_storage_size: Storage size should align with cluster size
- zookeeper_storage_size: Zookeeper storage should align with cluster size
Note: The small configuration is suitable for development and testing. Production environments typically require the large configuration for adequate performance and reliability.
kafka_storage_class¤
Storage class for Kafka broker persistent storage (must support ReadWriteOnce).
- Optional
- Environment Variable:
KAFKA_STORAGE_CLASS - Default Value: Auto-detected from available storage classes in cluster
Purpose: Specifies the storage class for Kafka broker persistent storage, which requires ReadWriteOnce (RWO) access mode. This is where Kafka stores message logs and data.
When to use: - Leave unset for automatic detection (recommended) - Set explicitly when you need a specific storage class - Must be a storage class supporting RWO access mode - Typically block-based storage for performance
Valid values: Any storage class name supporting ReadWriteOnce access mode
Impact: Kafka broker performance depends heavily on storage performance. Choose high-performance storage for production workloads. Incorrect storage class or one not supporting RWO will cause deployment to fail.
Related variables:
- kafka_storage_size: Size of storage for Kafka brokers
- zookeeper_storage_class: Separate storage class for Zookeeper
Note: Block-based storage classes typically provide better performance for Kafka (e.g., ibmc-block-gold, ocs-storagecluster-ceph-rbd). This variable only applies to Strimzi and Red Hat AMQ Streams providers.
kafka_storage_size¤
Size of persistent storage for Kafka brokers.
- Optional
- Environment Variable:
KAFKA_STORAGE_SIZE - Default Value:
100Gi
Purpose: Specifies the size of persistent volumes for Kafka broker storage. This is where Kafka stores all message logs and topic data.
When to use:
- Use default (100Gi) for development/test environments
- Increase significantly for production based on message volume and retention
- Plan for message throughput and retention period
- Monitor usage and expand as needed
Valid values: Kubernetes storage size format (e.g., 100Gi, 500Gi, 1Ti, 2Ti)
Impact: Insufficient storage will cause Kafka to stop accepting messages when volumes fill up. Size appropriately for your message volume and retention requirements.
Related variables:
- kafka_storage_class: Storage class for these volumes
- kafka_cluster_size: Larger clusters may need more storage
- Message retention settings affect storage requirements
Note: The default 100Gi is suitable for small deployments only. Production environments typically require 500Gi or more depending on message volume and retention. Plan storage based on daily message volume × retention days.
zookeeper_storage_class¤
Storage class for Zookeeper persistent storage (must support ReadWriteOnce).
- Optional
- Environment Variable:
ZOOKEEPER_STORAGE_CLASS - Default Value: Auto-detected from available storage classes in cluster
Purpose: Specifies the storage class for Zookeeper persistent storage, which requires ReadWriteOnce (RWO) access mode. Zookeeper stores cluster metadata and coordination data.
When to use:
- Leave unset for automatic detection (recommended)
- Set explicitly when you need a specific storage class
- Must be a storage class supporting RWO access mode
- Can be same as or different from kafka_storage_class
Valid values: Any storage class name supporting ReadWriteOnce access mode
Impact: Zookeeper is critical for Kafka cluster coordination. Reliable storage is essential. Incorrect storage class or one not supporting RWO will cause deployment to fail.
Related variables:
- zookeeper_storage_size: Size of storage for Zookeeper
- kafka_storage_class: Separate storage class for Kafka brokers
Note: Zookeeper storage requirements are typically much smaller than Kafka broker storage. This variable only applies to Strimzi and Red Hat AMQ Streams providers.
zookeeper_storage_size¤
Size of persistent storage for Zookeeper nodes.
- Optional
- Environment Variable:
ZOOKEEPER_STORAGE_SIZE - Default Value:
10Gi
Purpose: Specifies the size of persistent volumes for Zookeeper storage. Zookeeper stores cluster metadata, configuration, and coordination data.
When to use:
- Use default (10Gi) for most deployments
- Increase only for very large Kafka clusters with many topics/partitions
- Zookeeper storage needs are typically much smaller than Kafka broker storage
Valid values: Kubernetes storage size format (e.g., 10Gi, 20Gi, 50Gi)
Impact: Insufficient Zookeeper storage can cause cluster coordination issues. However, Zookeeper storage requirements are typically modest.
Related variables:
- zookeeper_storage_class: Storage class for these volumes
- kafka_storage_size: Kafka broker storage (typically much larger)
Note: The default 10Gi is sufficient for most deployments. Zookeeper storage requirements grow slowly with cluster size. Only increase if you have a very large number of topics and partitions.
kafka_user_name¤
Username for MAS to authenticate with Kafka.
- Optional
- Environment Variable:
KAFKA_USER_NAME - Default Value:
masuser
Purpose: Defines the Kafka user that will be created for MAS to authenticate with the Kafka cluster. This user is configured with appropriate permissions for MAS operations.
When to use:
- Use default (masuser) for standard MAS deployments
- Set to custom name when organizational policies require specific usernames
- Must be unique within the Kafka cluster
Valid values: Valid Kafka username string (e.g., masuser, mas-kafka-user)
Impact: This user is created in Kafka with necessary permissions and credentials are included in the generated KafkaCfg for MAS.
Related variables:
- kafka_user_password: Password for this user (Strimzi 0.25.0+/AMQ Streams 2.x+)
- Used in generated KafkaCfg when mas_instance_id is provided
Note: The default masuser is suitable for most deployments. This variable only applies to Strimzi and Red Hat AMQ Streams providers.
kafka_user_password¤
Password for the Kafka user (Strimzi 0.25.0+/AMQ Streams 2.x+).
- Optional
- Environment Variable:
KAFKA_USER_PASSWORD - Default Value: Randomly generated if not specified
Purpose: Sets the password for the Kafka user specified in kafka_user_name. This password is used for SCRAM-SHA authentication.
When to use: - Leave unset for automatic random password generation (recommended) - Set explicitly when you need a specific password - Requires Strimzi operator 0.25.0+ or AMQ Streams 2.x+
Valid values: Strong password string meeting security requirements
Impact: This password is stored in Kafka secrets and included in the generated KafkaCfg for MAS. If not set, a secure random password is generated automatically.
Related variables:
- kafka_user_name: User for which this password is set
- Used in generated KafkaCfg when mas_instance_id is provided
Note: Automatic password generation is recommended for security. This feature requires Strimzi operator version 0.25.0 or AMQ Streams 2.x+. Keep passwords secure and do not commit to source control.
mas_instance_id¤
MAS instance ID for generating KafkaCfg configuration.
- Optional
- Environment Variable:
MAS_INSTANCE_ID - Default Value: None
Purpose: Identifies which MAS instance the generated KafkaCfg will target. When set (along with mas_config_dir), the role generates a KafkaCfg resource for configuring MAS to use this Kafka cluster.
When to use:
- Set when you want to generate MAS configuration automatically
- Must match the instance ID of your MAS installation
- Required together with mas_config_dir for KafkaCfg generation
Valid values: Lowercase alphanumeric string, 3-12 characters (e.g., prod, dev, inst1)
Impact: When set with mas_config_dir, generates a KafkaCfg YAML file that can be applied to configure MAS. Without this, no MAS configuration is generated.
Related variables:
- mas_config_dir: Required together with this for KafkaCfg generation
- Generated KafkaCfg targets this MAS instance
Note: If either mas_instance_id or mas_config_dir is not set, the role will not generate a KafkaCfg template. You'll need to configure MAS manually.
mas_config_dir¤
Local directory path where generated KafkaCfg will be saved.
- Optional
- Environment Variable:
MAS_CONFIG_DIR - Default Value: None
Purpose: Specifies where to save the generated KafkaCfg YAML file. This file can be manually applied to configure MAS or used with the suite_config role for automated configuration.
When to use:
- Set when you want to generate MAS configuration automatically
- Use the same directory across all MAS setup roles for consistency
- Required together with mas_instance_id for KafkaCfg generation
Valid values: Any valid local filesystem path (e.g., /home/user/masconfig, ~/masconfig, ./config)
Impact: When set with mas_instance_id, generates kafkacfg-{provider}-system.yaml in this directory. The file can be applied with oc apply or used with suite_config role.
Related variables:
- mas_instance_id: Required together with this for KafkaCfg generation
- kafka_provider: Affects the generated filename
Note: If either mas_instance_id or mas_config_dir is not set, no KafkaCfg template is generated. Ensure the directory exists and is writable.
custom_labels¤
Comma-separated list of key=value labels to apply to Kafka resources.
- Optional
- Environment Variable:
CUSTOM_LABELS - Default Value: None
Purpose: Adds Kubernetes labels to Kafka-related resources for organization, selection, and filtering. Labels enable resource tracking, cost allocation, and custom automation.
When to use:
- Use to add organizational metadata (e.g., cost-center=engineering, environment=production)
- Use to enable resource tracking and cost allocation
- Use to support custom automation or monitoring tools
- Use to comply with organizational labeling standards
Valid values: Comma-separated list of key=value pairs (e.g., env=prod,team=platform,component=kafka)
Impact: Labels are applied to Kafka-related resources and can be used for filtering with oc get commands, monitoring queries, and automation scripts. Labels do not affect Kafka functionality.
Related variables: Works alongside Kubernetes resource labels for comprehensive resource management.
Note: This variable applies to Strimzi and Red Hat AMQ Streams providers. Labels help with resource organization and are especially useful in multi-tenant environments.
IBM Cloud Evenstreams Role Variables
ibmcloud_apikey¤
Defines IBM Cloud API Key. This API Key needs to have access to manage (provision/deprovision) IBM Cloud Event Streams.
- Required
- Environment Variable:
IBMCLOUD_APIKEY - Default Value: None
eventstreams_resourcegroup¤
Defines the IBM Cloud Resource Group to target the Event Streams instance.
- Optional
- Environment Variable:
EVENTSTREAMS_RESOURCEGROUP - Default Value:
Defaultor value defined byIBMCLOUD_RESOURCEGROUP
eventstreams_name¤
Event Streams instance name.
- Required
- Environment Variable:
EVENTSTREAMS_NAME - Default Value: None
eventstreams_plan¤
Event Streams instance plan.
- Optional
- Environment Variable:
EVENTSTREAMS_PLAN - Default Value:
standard
eventstreams_location¤
- Optional
- Environment Variable:
EVENTSTREAMS_LOCATION - Default Value:
us-eastor value defined byIBMCLOUD_REGION
eventstreams_retention¤
Event Streams topic retention period (in miliseconds).
- Optional
- Environment Variable:
EVENTSTREAMS_RETENTION - Default Value:
1209600000
eventstreams_create_manage_jms_topic¤
Defines whether to create specific Manage application JMS topics by default.
- Optional
- Environment Variable:
EVENTSTREAMS_CREATE_MANAGE_JMS_TOPICS - Default Value:
True
mas_instance_id¤
The instance ID of Maximo Application Suite that the KafkaCfg configuration will target. If this or mas_config_dir are not set then the role will not generate a KafkaCfg template.
- Environment Variable:
MAS_INSTANCE_ID - Default Value: None
mas_config_dir¤
Local directory to save the generated KafkaCfg resource definition. This can be used to manually configure a MAS instance to connect to the Kafka cluster, or used as an input to the suite_config role. If this or mas_instance_id are not set then the role will not generate a KafkaCfg template.
- Environment Variable:
MAS_CONFIG_DIR - Default Value: None
custom_labels¤
List of comma separated key=value pairs for setting custom labels on instance specific resources.
- Optional
- Environment Variable:
CUSTOM_LABELS - Default Value: None
Example Playbook¤
- hosts: localhost
any_errors_fatal: true
vars:
# Set storage class suitable for use on IBM Cloud ROKS
kafka_storage_class: ibmc-block-gold
# Generate a KafkaCfg template
mas_instance_id: masinst1
mas_config_dir: ~/masconfig
roles:
- ibm.mas_devops.kafka
AWS MSK Variables¤
Prerequisites¤
To run this role successfully you must have already installed the AWS CLI.
Also, you need to have AWS user credentials configured via aws configure command or simply export AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment variables with your corresponding AWS username credentials prior running this role.
kafka_version¤
The version of Kafka to deploy for AWS MSK.
- Environment Variable:
KAFKA_VERSION - Default Value:
3.3.1
kafka_cluster_name¤
The name of the Kafka cluster that will be created
- Required
- Environment Variable:
KAFKA_CLUSTER_NAME - Default Value:
maskafka
aws_region¤
- Required
- Environment Variable:
AWS_REGION - Default Value: None
vpc_id¤
The AWS Virtual Private Cloud identifier (VPC ID) where the MSK instance will be hosted.
- Required
- Environment Variable:
VPC_ID - Default Value: None
aws_msk_cidr_az1¤
The CIDR address for the first Availability Zone subnet. This information is found in the subnet details under your VPC.
- Required
- Environment Variable:
AWS_MSK_CIDR_AZ1 - Default Value: None
aws_msk_cidr_az2¤
The CIDR address for the second Availability Zone subnet. This information is found in the subnet details under your VPC.
- Required
- Environment Variable:
AWS_MSK_CIDR_AZ2 - Default Value: None
aws_msk_cidr_az3¤
The CIDR address for the third Availability Zone subnet. This information is found in the subnet details under your VPC.
- Required
- Environment Variable:
AWS_MSK_CIDR_AZ3 - Default Value: None
aws_msk_ingress_cidr¤
The IPv4 CIDR address for ingress connection. This information is found in the subnet details under your VPC.
- Required
- Environment Variable:
AWS_MSK_INGRESS_CIDR - Default Value: None
aws_msk_egress_cidr¤
The IPv4 CIDR address for egress connection. This information is found in the subnet details under your VPC.
- Required
- Environment Variable:
AWS_MSK_EGRESS_CIDR - Default Value: None
aws_kafka_user_name¤
The name of the user to setup in the cluster for MAS.
- Required
- Environment Variable:
AWS_KAFKA_USER_NAME - Default Value: None
aws_kafka_user_password¤
The password of the user to setup in the cluster for MAS.
- Optional
- Environment Variable:
AWS_KAFKA_USER_PASSWORD - Default Value: None
aws_msk_instance_type¤
The type/flavor of your MSK instance.
- Optional
- Environment Variable:
AWS_MSK_INSTANCE_TYPE - Default Value:
kafka.m5.large
aws_msk_volume_size¤
The storage/volume size of your MSK instance.
- Optional
- Environment Variable:
AWS_MSK_VOLUME_SIZE - Default Value:
100
aws_msk_instance_number¤
The number of broker/instances of your MSK instance.
- Optional
- Environment Variable:
AWS_MSK_INSTANCE_NUMBER - Default Value:
3
mas_instance_id¤
The instance ID of Maximo Application Suite that the KafkaCfg configuration will target. If this or mas_config_dir are not set then the role will not generate a KafkaCfg template.
- Environment Variable:
MAS_INSTANCE_ID - Default Value: None
mas_config_dir¤
Local directory to save the generated KafkaCfg resource definition. This can be used to manually configure a MAS instance to connect to the Kafka cluster, or used as an input to the suite_config role. If this or mas_instance_id are not set then the role will not generate a KafkaCfg template.
- Environment Variable:
MAS_CONFIG_DIR - Default Value: None
custom_labels¤
List of comma separated key=value pairs for setting custom labels on instance specific resources.
- Optional
- Environment Variable:
CUSTOM_LABELS - Default Value: None
aws_msk_secret¤
AWS MSK Secret name. The secret name must begin with the prefix AmazonMSK_. If this is not set, then default secret name will be AmazonMSK_SECRET_{{kafka_cluster_name}}
- Optional
- Environment Variable:
AWS_MSK_SECRET - Default Value:
AmazonMSK_SECRET_{{kafka_cluster_name}}'
Install AWS MSK¤
- hosts: localhost
any_errors_fatal: true
vars:
aws_region: ca-central-1
aws_access_key_id: *****
aws_secret_access_key: *****
kafka_version: 3.3.1
kafka_provider: aws
kafka_action: install
kafka_cluster_name: msk-abcd0zyxw
kafka_namespace: msk-abcd0zyxw
vpc_id: vpc-07088da510b3c35c5
aws_kafka_user_name: mskuser-abcd0zyxw
aws_msk_instance_type: kafka.t3.small
aws_msk_volume_size: 100
aws_msk_instance_number: 3
aws_msk_cidr_az1: "10.0.128.0/20"
aws_msk_cidr_az2: "10.0.144.0/20"
aws_msk_cidr_az3: "10.0.160.0/20"
aws_msk_ingress_cidr: "10.0.0.0/16"
aws_msk_egress_cidr: "10.0.0.0/16"
# Generate a KafkaCfg template
mas_config_dir: /var/tmp/masconfigdir
mas_instance_id: abcd0zyxw
roles:
- ibm.mas_devops.kafka
Uninstall AWS MSK¤
- hosts: localhost
any_errors_fatal: true
vars:
aws_region: ca-central-1
aws_access_key_id: *****
aws_secret_access_key: *****
vpc_id: vpc-07088da510b3c35c5
kafka_provider: aws
kafka_action: uninstall
kafka_cluster_name: msk-abcd0zyxw
aws_msk_cidr_az1: "10.0.128.0/20"
aws_msk_cidr_az2: "10.0.144.0/20"
aws_msk_cidr_az3: "10.0.160.0/20"
roles:
- ibm.mas_devops.kafka
Run Role Playbook¤
export KAFKA_STORAGE_CLASS=ibmc-block-gold
export MAS_INSTANCE_ID=masinst1
export MAS_CONFIG_DIR=~/masconfig
ansible-playbook ibm.mas_devops.run_role
License¤
EPL-2.0