What is the role of a Kafka consumer?
Publish messages to topics
Manage the cluster and its configuration
Store message offsets
Subscribe to and process messages from topics
What is a 'Partition' in the context of Kafka topics?
A configuration setting that determines message retention time.
A security mechanism to control access to specific topics.
A logical grouping of messages within a topic for parallel processing.
A separate copy of the entire topic data for redundancy.
In Kafka, what is a 'Topic' analogous to?
A specific partition within a broker
A category or stream of messages
A storage location on a hard drive
A single message
What does it mean for a Kafka message to be 'persisted'?
It is encrypted for security purposes
It is stored in memory for fast access
It is immediately deleted after being read
It is written to disk and stored durably
What is the primary function of a Kafka producer?
Reading messages from a topic
Storing messages persistently
Managing user permissions
Publishing messages to a topic
How are messages within a Kafka topic organized?
Sorted by message key
Divided into ordered, immutable sequences called partitions
Grouped by consumer group
Randomly distributed
How are messages within a Kafka topic further divided for scalability and parallelism?
Segments
Partitions
Offsets
Clusters
What is the purpose of the auto.create.topics.enable configuration in Kafka?
auto.create.topics.enable
Configures automatic data balancing across partitions within a topic.
Automatically creates topics if they don't exist when producers or consumers try to access them.
Enables automatic topic deletion after a specified retention period.
Allows dynamic adjustment of the replication factor for existing topics.
Which Kafka CLI tool is used to produce messages to a Kafka topic?
kafka-topics.sh
kafka-console-consumer.sh
kafka-configs.sh
kafka-console-producer.sh
How does Kafka achieve scalability?
By limiting the number of topics allowed
Through complex data compression algorithms
By using a single, powerful server
By distributing data across multiple brokers