How can load balancing of Kafka clients across multiple brokers be accomplished?
Answer : A
Partitions are the primary mechanism for achieving load balancing in Kafka. When a topic has multiple partitions, Kafka clients (producers and consumers) can distribute the load across brokers hosting these partitions.
Which technology can be used to perform event stream processing? (Choose two.)
Answer : B, C
Kafka Streams is a client library for building real-time applications that process and analyze data stored in Kafka.
ksqlDB enables event stream processing using SQL-like queries, allowing real-time transformation and analysis of Kafka topics.
Which valid security protocols are included for broker listeners? (Choose three.)
Answer : A, B, D
What is the primary purpose of Kafka quotas?
Answer : A
Kafka quotas are used to limit the throughput (bytes/sec) of producers and consumers to ensure fair resource usage and prevent any single client from overwhelming the brokers.
Kafka Connect is running on a two node cluster in distributed mode. The connector is a source connector that pulls data from Postgres tables (users/payment/orders), writes to topics with two partitions, and with replication factor two. The development team notices that the data is lagging behind.
What should be done to reduce the data lag*?
The Connector definition is listed below:
{
"name": "confluent-postgresql-source",
"connector class": "PostgresSource",
"topic.prefix": "postgresql_",
& nbsp;& nbsp;& nbsp;...
"db.name": "postgres",
"table.whitelist": "users.payment.orders'',
"timestamp.column.name": "created_at",
"output.data format": "JSON",
"db.timezone": "UTC",
"tasks.max": "1"
}
Answer : B
The connector is currently configured with 'tasks.max': '1', which means only one task is handling all tables (users, payment, orders). This can create a bottleneck and lead to lag. Increasing tasks.max allows Kafka Connect to parallelize work across multiple tasks, which can pull data from different tables concurrently and reduce lag.
A Kafka cluster with three brokers has a topic with 10 partitions and a replication factor set to three. Each partition stores 25 GB data per day and data retention is set to 24 hours.
How much storage will be consumed by the topic on each broker?
Answer : C
10 partitions 25 GB/day = 250 GB total per day for the topic (primary data).
With a replication factor of 3, there are 3 full copies of the data: 250 GB 3 = 750 GB total across the entire cluster.
The cluster has 3 brokers, and Kafka tries to distribute replicas evenly among them: 750 GB 3 brokers = 250 GB per broker on average.
However, due to replication, some partitions have leaders and followers, so there's some overlap and not-perfect distribution. Each broker stores approximately 2/3 of the total topic data (since each broker holds replicas for around 2/3 of the partitions).
2/3 750 GB = 500 GB, but this is shared, so each broker ends up storing ~300 GB of replicated data, including its share of leaders and followers.
A broker in the Kafka cluster is currently acting as the Controller.
Which statement is correct?
Answer : A
The Controller broker is a regular broker that also takes on additional responsibilities for managing cluster metadata, such as leader elections and partition assignments. It still hosts topic partitions and participates in replication like any other broker.