Skip to content

Kafka

Changelog (last updated v2.2)
v2.2: single-writer support — two topics per database

Single-writer indexing requires two Kafka topics per database: a source log for client writes and a replica log for the indexing leader’s resolved output. XTDB uses Kafka’s consumer-group rebalance protocol to elect the leader for each database, and a transactional producer per leader to fence split-brain writes to the replica log — see ‘Leader election and fencing’ below.

Previously, a database used a single Kafka topic, and every indexer node consumed it independently. With single-writer, only the elected leader consumes the source topic; followers tail the replica topic instead.

Upgrading:

  • The replica topic defaults to ${topic}-replica and auto-creates when autoCreateTopic is enabled, so existing deployments need no configuration changes to pick it up.
  • If multiple XTDB deployments share a Kafka cluster, set transactionalIdPrefix on each deployment’s !Kafka cluster config — otherwise their leaders will fence each other across deployment boundaries.
  • ACL-restricted topics now need Describe / Read / Write on both the source and replica topics, plus transactional-producer permissions (see Setup).
v2.1: multi-database support

As part of multi-database support, logClusters were extracted in v2.1.

Prior to that, the configuration in logClusters was within the log:

log: !Kafka
bootstrapServers: "localhost:9092"
topic: "xtdb-log"
# autoCreateTopic: true
# pollDuration: "PT1S"
# propertiesFile: "kafka.properties"
# propertiesMap:
# became
logClusters:
kafkaCluster: !Kafka
bootstrapServers: "localhost:9092"
# pollDuration: "PT1S"
# propertiesFile: "kafka.properties"
# propertiesMap:
log: !Kafka
cluster: kafkaCluster
topic: "xtdb-log"
# autoCreateTopic: true

Apache Kafka can be used as XTDB’s message log. Each database uses two Kafka topics — a source log for client writes and a replica log for the indexing leader’s resolved output — plus Kafka’s consumer-group protocol to elect the leader for that database automatically. See ‘Database architecture’ for the concepts; this page covers how to set Kafka up to back them.

  1. Add a dependency to the com.xtdb/xtdb-kafka module in your dependency manager.

  2. On your Kafka cluster, XTDB requires two topics per database — a source log and a replica log:

    • Both can be created manually and provided to the node config, or XTDB can create them automatically.
    • If allowing XTDB to create the topics automatically, ensure that the connection properties supplied to the XTDB node have the appropriate permissions to create topics — XTDB will create each with the expected configuration values (single partition, LogAppendTime timestamps).
  3. Both topics should be configured with the following properties:

    • A single partition — this ensures that the log is strictly ordered, and single-writer leader election (see below) relies on Kafka assigning the single partition to a single consumer at a time.
    • message.timestamp.type set to LogAppendTime — ensures the timestamp of the message is the time it was appended to the log, rather than the time it was sent by the producer.
    • The XTDB log is generally set up using the default values for configuration. A few key values to consider:
      • retention.ms: as messages are not required to be permanently on the log, this value does not need to be particularly high. The default value of 1 day is sufficient for most use cases. When extra precaution should be taken to prevent any data loss on certain environments, a larger value is recommended, with 1 week as a starting point.

      • max.message.bytes: generally, this is using the default value of 1MB, which is fit for purpose for most log messages. This will depend on the overall size of transactions that are being sent into XTDB.

      • cleanup.policy: The Kafka module within XTDB does not make use of compacted messages, so it is recommended that the topic cleanup policy should use the default value of delete.

  4. XTDB should be configured to use the topics, and the Kafka cluster they’re hosted on. It should also be authorised to perform all of the necessary operations on both.

    • For configuring the Kafka module to authenticate with the Kafka cluster, use the propertiesFile or propertiesMap configuration options to supply the necessary connection properties. See the example configuration below.
    • If the Kafka cluster is using ACLs, the XTDB node needs:
      • Describe / Read / Write on both the source and replica topics.
      • Describe / Write on the TransactionalId resource matching the replica-log producer’s transactional ID (defaults to ${databaseName}-leader, or ${transactionalIdPrefix}-${databaseName}-leader when a prefix is set — see Leader election and fencing).

To use the Kafka module, include the following in your node configuration:

## We first declare the Kafka log cluster:
logClusters:
# You can define multiple Kafka clusters here, and refer to them by name in the log configuration.
# Here we define a single Kafka cluster named "kafkaCluster".
kafkaCluster: !Kafka
# -- required
# A comma-separated list of host:port pairs to use for establishing the
# initial connection to the Kafka cluster.
# (Can be set as an !Env value)
bootstrapServers: "localhost:9092"
# -- optional
# The maximum time to block waiting for records to be returned by the Kafka consumer.
# pollDuration: "PT1S"
# Path to a Java properties file containing Kafka connection properties,
# supplied directly to the Kafka client.
# (Can be set as an !Env value)
# propertiesFile: "kafka.properties"
# A map of Kafka connection properties, supplied directly to the Kafka client.
# propertiesMap:
# Prefix applied to the transactional IDs used by replica-log producers (v2.2+).
# Required when multiple XTDB deployments share a Kafka cluster, to avoid them
# fencing each other's leaders. Leave unset otherwise.
# transactionalIdPrefix: "prod"
# Consumer-group ID used for per-database leader election (v2.2+).
# Defaults to "xtdb" — change it only if you need a separate group per deployment
# sharing a Kafka cluster (in most cases, `transactionalIdPrefix` is sufficient).
# groupId: "xtdb"
## For the database, we then create a log using the Kafka cluster we just defined:
log: !Kafka
# -- required
# The name of the Kafka cluster to use for the source log.
cluster: kafkaCluster
# Name of the Kafka topic to use for the source log.
# (Can be set as an !Env value)
topic: "xtdb-log"
# -- optional
# The name of the Kafka cluster to use for the replica log (v2.2+).
# Defaults to the same cluster as the source log.
# replicaCluster: kafkaCluster
# Name of the Kafka topic to use for the replica log (v2.2+).
# Defaults to "${topic}-replica".
# replicaTopic: "xtdb-log-replica"
# Whether or not to automatically create the topics, if they do not already exist.
# Applies to both the source and replica topics.
# autoCreateTopic: true

The following piece of node configuration demonstrates the following common use case:

  • Cluster is secured with SASL - authentication is required from the module.
  • Topic has already been created manually.
  • Configuration values are being passed in as environment variables.
logClusters:
kafkaCluster: !Kafka
bootstrapServers: !Env KAFKA_BOOTSTRAP_SERVERS
propertiesMap:
sasl.mechanism: PLAIN
security.protocol: SASL_SSL
sasl.jaas.config: !Env KAFKA_SASL_JAAS_CONFIG
log: !Kafka
topic: !Env XTDB_LOG_TOPIC
autoCreateTopic: false

The KAFKA_SASL_JAAS_CONFIG environment variable will likely contain a string similar to the following, and should be passed in as a secret value:

org.apache.kafka.common.security.plain.PlainLoginModule required username="username" password="password";

The ‘Database architecture’ page describes XTDB’s single-writer indexing model in terms of properties — exactly one leader per database, automatic failover, followers as hot standbys. This section describes how the Kafka module actually enforces those properties.

Every XTDB node running the Kafka log subscribes to each database’s source topic via a shared Kafka consumer group (groupId, defaulting to "xtdb"). Kafka’s rebalance protocol then assigns each topic’s single partition to exactly one consumer across the group — that consumer is the leader for that database. When a leader node stops (crash, network partition, long GC pause) or a new node joins the group, Kafka triggers a rebalance and the assignment moves.

Because all databases on a node share one consumer group via a single underlying consumer, Kafka’s CooperativeStickyAssignor distributes leaderships evenly across the cluster — e.g. with three nodes serving three databases, each node ends up leader for one database and follower for the other two.

Once elected, a leader writes to its database’s replica log via a Kafka transactional producer with a fixed transactional ID: ${databaseName}-leader (or ${transactionalIdPrefix}-${databaseName}-leader when a prefix is set).

If a new leader is elected and opens a producer with the same transactional ID, Kafka’s transaction coordinator fences the previous one — any in-flight writes from the outgoing leader raise ProducerFencedException and fail cleanly. This guarantees at most one node is ever writing to the replica log for a given database, even across unclean handovers.

Sharing a Kafka cluster across deployments

Section titled “Sharing a Kafka cluster across deployments”

Kafka’s transaction coordinator keys transactional IDs globally, not per-topic. If you run multiple XTDB deployments against the same Kafka cluster (e.g. staging + prod, or multiple tenants), their leaders will fence each other because they share transactional IDs like xtdb-leader.

Set transactionalIdPrefix on each deployment’s cluster config to disambiguate:

logClusters:
kafkaCluster: !Kafka
bootstrapServers: "localhost:9092"
transactionalIdPrefix: "prod" # → "prod-xtdb-leader", etc.

All logs on a cluster (source/replica for every database, including secondaries attached at runtime) inherit the prefix automatically.

Kafka-backed logs offer strong durability, but require tuning and backup strategies to align with your recovery objectives.

To minimize the risk of data loss:

  • Replicate the topic - set a replication factor of 3+ for fault tolerance
  • Enforce quorum writes - use min.insync.replicas > 1
  • Tune retention - ensure retention.ms and/or retention.bytes keep unindexed messages long enough to allow for safe backup or flushing

XTDB sets safe producer defaults, but you must verify your topic-level configs.

See Apache Kafka documentation for details.

Managed services like Confluent Cloud may offer higher guarantees and simplified observability.

There are three main ways to safeguard your XTDB Kafka log:

Use Kafka-native tools to replicate log data between clusters:

This allows for:

  • Geo-redundancy
  • Low-RPO disaster recovery
  • Hot-standby clusters

Note: Replication does not replace backups --- it only increases availability.

XTDB can rebuild its state from upstream sources (event logs, message queues) used to submit transactions.

Advantages:

  • Independent recovery source
  • Replay can be filtered, transformed, or validated
  • Fills gaps between backup and failure