Configuring Apache Kafka topics output from BVR

The following, if enabled by configuration, will produce data for an associated kafka topic:

Table 1. Apache Kafka topics
Description Topic Name BVR Configuration details
Call and Conference Detail Records (CDRs) output by BVR blueworx.cdr CDR Configuration Options
Application logging (high level application events and user generated logging) blueworx.application Log Configuration Options
Errors, Warnings and emergency Informational messages blueworx.error Log Configuration Options
Low level tracing blueworx.trace Log Configuration Options

Before using Apache Kafka to process the data streams that you have configured in BVR you must create the topics in Apache Kafka that you wish to use.

Creating Apache Kafka topics

The Apache Kafka Quickstart guide details how to create a simple topic, using broker defaults. An example for each supported topic follows:

bin/ --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic blueworx.cdr
bin/ --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic blueworx.application
bin/ --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic blueworx.error
bin/ --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic blueworx.trace

Additional topic parameters can be specified to override the server defaults. These parameters are detailed in the Apache Kafka documentation - Topic-level configs.

Retention policies

When creating a topic consider the topic retention policy. The topic parameters will determine how long segments (messages) are kept in a topic (, whether segments are deleted or compressed (cleanup.policy), at what interval the deleted messages are flushed (, etc.

When deciding on a retention policy, consider if the data written to the topic is critical to your organization, and if it is you must be sure that the required Apache Kafka consumers are able to read the data from the topic before it is deleted or you will lose the data. You may want to alter the retention policy to ensure this and implement mechanisms to manage consumers used to read the data.

Multi-broker cluster configuration

To ensure high availability in your Apache Kafka environment it is recommended that you set up a multi-broker Kafka cluster, as detailed in Installing Apache Kafka (optional). When operating in a multi-broker cluster you will need to change the replication-factor parameter. The Apache Kafka documentation states:

"The replication factor controls how many servers will replicate each message that is written. If you have a replication factor of 3 then up to 2 servers can fail before you will lose access to your data. We [Apache Kafka project] recommend you use a replication factor of 2 or 3 so that you can transparently bounce machines without interrupting data consumption."

See Apache Kafka :: Adding and removing topics for more details.

Modifying topic configuration

If you need to modify the topic configuration, see Apache Kafka :: Modifying topics in the Apache Kafka documentation.