What is the role of a 'state store' in Kafka Streams?
To cache data from external databases.
To persist data required for stateful operations.
To store intermediate results of stateless operations.
To buffer incoming records before processing.
What is the purpose of using tumbling windows in Kafka Streams?
To process records in fixed-size, non-overlapping time intervals.
To divide data into sessions based on user activity.
To trigger aggregations only when a specific event occurs.
To overlap windows for smoothing out aggregated results.
What is the significance of 'Exactly Once Semantics' in Kafka Streams?
It prioritizes speed over accuracy in data processing.
It ensures that records are processed in the exact order they were produced.
It prevents duplicate processing of records even in the event of failures.
It guarantees that each record is processed at least once.
How does Kafka Streams achieve fault tolerance?
By replicating stream processing tasks across multiple nodes.
By relying solely on message acknowledgments from consumers.
By using a single, centralized processing unit.
By storing all processed data in a separate, redundant database.
What is the role of a Kafka Controller in a cluster?
Performing load balancing of messages
Managing message consumption rates
Monitoring and managing the health of brokers
Handling data replication between brokers
How does Grafana complement Prometheus in Kafka monitoring?
It sends alerts based on Kafka metrics collected by JMX
It provides customizable dashboards and visualizations for Kafka metrics stored in Prometheus
It acts as a JMX client to collect metrics from Kafka brokers
It replaces Prometheus as the primary metric storage
What is a key difference between Source Connectors and Sink Connectors in Kafka Connect?
Source Connectors require custom coding, while Sink Connectors use pre-built configurations.
Source Connectors push data, while Sink Connectors pull data.
Source Connectors are stateful, while Sink Connectors are stateless.
Source Connectors handle real-time data, while Sink Connectors handle batch data.
What is the primary benefit of using Kafka's idempotent producer feature?
Elimination of duplicate messages on the broker due to producer retries.
Automatic data balancing across multiple Kafka brokers.
Improved message ordering guarantees within a partition.
Increased throughput by reducing the need for message acknowledgments.
What is the primary purpose of log compaction in Kafka?
Optimizing message routing
Retaining the latest value for each key
Improving message compression
Deleting old messages based on time
Which of the following is a common format for specifying connector configurations in Kafka Connect?
YAML
All of the above
JSON
Properties files