Pre-Summer Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmas50

Confluent CCDAK - Confluent Certified Developer for Apache Kafka Certification Examination

Page: 1 / 3
Total 90 questions

You need to consume messages from Kafka using the command-line interface (CLI).

Which command should you use?

A.

kafka-console-consumer

B.

kafka-consumer

C.

kafka-get-messages

D.

kafka-consume

(Which configuration is valid for deploying a JDBC Source Connector to read all rows from the orders table and write them to the dbl-orders topic?)

A.

{"name": "orders-connect","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl","topic.whitelist": "orders","auto.create": "true"}

B.

{"name": "dbl-orders","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.blacklist": "ord*"}

C.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&useAutoAuth=true","topic.prefix": "dbl-","table.whitelist": "orders"}

D.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.JdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.whitelist": "orders"}

The producer code below features a Callback class with a method called onCompletion().

When will the onCompletion() method be invoked?

A.

When a consumer sends an acknowledgement to the producer

B.

When the producer puts the message into its socket buffer

C.

When the producer batches the message

D.

When the producer receives the acknowledgment from the broker

Clients that connect to a Kafka cluster are required to specify one or more brokers in the bootstrap.servers parameter.

What is the primary advantage of specifying more than one broker?

A.

It provides redundancy in making the initial connection to the Kafka cluster.

B.

It forces clients to enumerate every single broker in the cluster.

C.

It is the mechanism to distribute a topic’s partitions across multiple brokers.

D.

It provides the ability to wake up dormant brokers.

Match each configuration parameter with the correct deployment step in installing a Kafka connector.

Which two statements are correct about transactions in Kafka?

(Select two.)

A.

All messages from a failed transaction will be deleted from a Kafka topic.

B.

Transactions are only possible when writing messages to a topic with single partition.

C.

Consumers can consume both committed and uncommitted transactions.

D.

Information about producers and their transactions is stored in the _transaction_state topic.

E.

Transactions guarantee at least once delivery of messages.

You have a topic t1 with six partitions. You use Kafka Connect to send data from topic t1 in your Kafka cluster to Amazon S3. Kafka Connect is configured for two tasks.

How many partitions will each task process?

A.

2

B.

3

C.

6

D.

12

(A stream processing application tracks user activity in online shopping carts, including items added, removed, and ordered throughout the day for each user.

You need to capture data to identify possible periods of user inactivity.

Which type of Kafka Streams window should you use?)

A.

Session

B.

Hopping

C.

Tumbling

D.

Sliding

Which is true about topic compaction?

A.

When a client produces a new event with an existing key, the old value is overwritten with the new value in the compacted log segment.

B.

When a client produces a new event with an existing key, the broker immediately deletes the offset of the existing event.

C.

Topic compaction does not remove old events; instead, when clients consume events from a compacted topic, they store events in a hashmap that maintains the latest value.

D.

Compaction will keep exactly one message per key after compaction of inactive log segments.

(You are developing a Java application that includes a Kafka consumer.

You need to integrate Kafka client logs with your own application logs.

Your application is using the Log4j2 logging framework.

Which Java library dependency must you include in your project?)

A.

SLF4J implementation for Log4j2 (org.apache.logging.log4j:log4j-slf4j-impl)

B.

SLF4J implementation for Log4j 1.2 (org.slf4j:slf4j-log4j12)

C.

Just the Log4j2 dependency of the application

D.

None, the correct dependency will be added transitively by the Kafka client