0% found this document useful (0 votes)
8 views5 pages

Chapter 7 Flume and Kafka Questions Answers

The document contains multiple-choice, single-choice, and true/false questions related to Flume and Kafka architectures, their components, features, and functionalities. It covers topics such as data sources, sinks, messaging patterns, reliability mechanisms, and the roles of various components. The questions are designed to test knowledge on log collection, message processing, and data handling in these systems.

Uploaded by

Mahmoud Ibrahim
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views5 pages

Chapter 7 Flume and Kafka Questions Answers

The document contains multiple-choice, single-choice, and true/false questions related to Flume and Kafka architectures, their components, features, and functionalities. It covers topics such as data sources, sinks, messaging patterns, reliability mechanisms, and the roles of various components. The questions are designed to test knowledge on log collection, message processing, and data handling in these systems.

Uploaded by

Mahmoud Ibrahim
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

Multiple-Choice Questions (Select multiple correct answers):

1. Which of the following are components of the Flume architecture?


○ A. Source
○ B. Channel
○ C. Sink
○ D. Interceptor
2. Which features are supported by Flume for log collection?
○ A. Custom data collection tasks
○ B. Data encryption and compression
○ C. Load balancing
○ D. High throughput
3. What types of data sources can Flume collect from?
○ A. Local files
○ B. REST messages
○ C. Kafka
○ D. Syslog
4. What are the characteristics of Flume's channel component?
○ A. Temporary data storage
○ B. Full transaction support
○ C. Provides ordering guarantees
○ D. Fault-tolerant
5. Which types of sinks are supported by Flume?
○ A. HDFS sink
○ B. Avro sink
○ C. Kafka sink
○ D. MorphlineSolr sink
6. Which of the following are key features of Kafka?
○ A. High throughput
○ B. Horizontal scaling
○ C. Message persistence
○ D. Real-time and offline processing
7. Which components are part of Kafka’s architecture?
○ A. Producer
○ B. Broker
○ C. ZooKeeper
○ D. Consumer
8. What are the two main messaging patterns supported by Kafka?
○ A. Point-to-point messaging
○ B. Publish-subscribe messaging
○ C. Stream-based messaging
○ D. Batch messaging
9. What are the advantages of Kafka's message partitioning?
○ A. Improves throughput
○ B. Ensures sequential message processing
○ C. Supports distributed consumption
○ D. Optimizes resource usage
10. What types of data storage options are supported by Kafka for reliability?
○ A. Data replication
○ B. Message offsets
○ C. High watermark tracking
○ D. Leader-follower replication
11. Which of the following describe Kafka’s reliability mechanisms?
○ A. Transaction management
○ B. Acknowledgment (ACK) mechanisms
○ C. Exactly-once processing
○ D. Idempotency
12. What are some uses of Kafka in real-world applications?
○ A. Log collection
○ B. Messaging systems
○ C. Stream processing
○ D. Batch processing
13. Which of the following describe Flume’s multi-agent architecture?
○ A. Cascading agents for log collection
○ B. Consolidating logs from multiple sources
○ C. Distributing logs to various sinks
○ D. Supporting real-time data replication
14. What are the primary components of a Kafka message?
○ A. Key
○ B. Value
○ C. Offset
○ D. Timestamp
15. Which of the following methods can Kafka use to handle old data?
○ A. Log deletion
○ B. Log compaction
○ C. Data archiving
○ D. Data replication

Single-Choice Questions (Select one correct answer):

1. What is the basic unit of data in Flume?


○ A. Event
○ B. Record
○ C. Log
○ D. Transaction
2. Which component in Kafka is responsible for managing distributed configurations
and leader elections?
○ A. Producer
○ B. ZooKeeper
○ C. Broker
○ D. Consumer
3. What is the primary purpose of Flume’s sink?
○ A. To collect log data
○ B. To store log data temporarily
○ C. To deliver data to a destination
○ D. To process log data
4. What type of messaging pattern does Kafka use to allow multiple consumers to
read the same message?
○ A. Point-to-point messaging
○ B. Publish-subscribe messaging
○ C. Queue-based messaging
○ D. Multicast messaging
5. What ensures Kafka’s fault tolerance?
○ A. Message ordering
○ B. Data replication
○ C. Consumer offsets
○ D. Partitioning
6. What is the function of Kafka’s producer?
○ A. To send messages to brokers
○ B. To consume messages from brokers
○ C. To monitor brokers
○ D. To manage partitioning
7. In Flume, what role does the channel play?
○ A. A source for data collection
○ B. A temporary buffer between source and sink
○ C. A final destination for data
○ D. A filter for log data
8. What is Kafka’s leader in the partition replication model?
○ A. The consumer that reads data
○ B. The broker that writes data
○ C. The replica responsible for all read and write requests
○ D. The backup replica for failover
9. Which sink would be used if Flume needs to send data to Hadoop?
○ A. Kafka sink
○ B. Avro sink
○ C. HDFS sink
○ D. Thrift sink
10. What mechanism does Kafka use to track consumer reading positions?
○ A. Offsets
○ B. Checkpoints
○ C. Partitions
○ D. Logs
11. What type of Kafka message delivery guarantees that a message will be delivered
only once?
○ A. At-most-once
○ B. At-least-once
○ C. Exactly-once
○ D. No guarantee
12. Which Kafka component is responsible for pulling messages from brokers?
○ A. Producer
○ B. Consumer
○ C. Broker
○ D. ZooKeeper
13. What ensures that a Flume sink only consumes events successfully sent to the
destination?
○ A. Transactions
○ B. Data replication
○ C. Failover
○ D. Compression
14. What is the main benefit of Kafka’s log compaction feature?
○ A. To store older messages for longer periods
○ B. To ensure the latest message version is retained
○ C. To delete all messages after consumption
○ D. To enhance message throughput
15. What type of Flume source runs a command or script to generate events?
○ A. Exec source
○ B. JMS source
○ C. Spooling directory source
○ D. HTTP source

True/False Questions:

1. Flume supports both driver-based and polling sources for data collection.
○ True/False
2. Kafka’s partitioning ensures that messages are consumed sequentially across all
partitions.
○ True/False (Each partition maintains its own order, but there is no guarantee
across partitions.)
3. In Kafka, a message can be consumed by more than one consumer if they are in
different consumer groups.
○ True/False
4. Flume channels are fully transactional, meaning data is either completely
processed or rolled back.
○ True/False
5. Kafka’s data replication only involves copying data from one partition to another.
○ True/False (Replication is done across brokers to ensure fault tolerance.)
6. A Flume agent can have multiple sources but only one sink.
○ True/False (Flume agents can have multiple sources and sinks.)
7. Kafka consumers pull messages from brokers, while producers push messages to
brokers.
○ True/False
8. Flume can only send log data to HDFS and HBase.
○ True/False (Flume can send data to various sinks including Kafka, Solr, and
others.)
9. In Kafka, message offsets are automatically reset when a consumer joins a new
group.
○ True/False (Offsets can be tracked and stored to avoid resetting.)
10. Flume's file channel offers higher performance but less reliability than the
memory channel.
○ True/False (The memory channel offers higher performance but less reliability.)
11. Kafka topics are used to categorize messages within the cluster.
○ True/False
12. Flume can filter and clean data during its transmission from source to sink.
○ True/False
13. Kafka’s leader-follower model ensures that consumers only interact with the
leader replica.
○ True/False
14. In Flume, an interceptor modifies data based on custom configurations before
sending it to the sink.
○ True/False
15. Kafka can guarantee exactly-once processing for all messages in all scenarios.
○ True/False (Exactly-once processing is complex and may not be guaranteed in
all cases.)

You might also like