Spring Sale - 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: dm70dm

CCDAK Confluent Certified Developer for Apache Kafka Certification Examination Questions and Answers

Questions 4

You have a Kafka client application that has real-time processing requirements.

Which Kafka metric should you monitor?

Options:

A.

Consumer lag between brokers and consumers

B.

Total time to serve requests to replica followers

C.

Consumer heartbeat rate to group coordinator

D.

Aggregate incoming byte rate

Buy Now
Questions 5

You create a topic named stream-logs with:

A replication factor of 3

Four partitions

Messages that are plain logs without a keyHow will messages be distributed across partitions?

Options:

A.

The first message will always be written to partition 0.

B.

Messages will be distributed round-robin among all the topic partitions.

C.

All messages will be written to the same log segment.

D.

Messages will be distributed among all the topic partitions with strict ordering.

Buy Now
Questions 6

Your Kafka cluster has five brokers. The topic t1 on the cluster has:

Two partitions

Replication factor = 4

min.insync.replicas = 3You need strong durability guarantees for messages written to topic t1.You configure a producer acks=all and all the replicas for t1 are in-sync.How many brokers need to acknowledge a message before it is considered committed?

Options:

A.

2

B.

3

C.

4

D.

5

Buy Now
Questions 7

You are writing a producer application and need to ensure proper delivery. You configure the producer with acks=all.

Which two actions should you take to ensure proper error handling?

(Select two.)

Options:

A.

Use a callback argument in producer.send() where you check delivery status.

B.

Check that producer.send() returned a RecordMetadata object and is not null.

C.

Surround the call of producer.send() with a try/catch block to catch KafkaException.

D.

Check the value of ProducerRecord.status().

Buy Now
Questions 8

You are composing a REST request to create a new connector in a running Connect cluster. You invoke POST /connectors with a configuration and receive a 409 (Conflict) response.

What are two reasons for this response? (Select two.)

Options:

A.

The connector configuration was invalid, and the response body will expand on the configuration error.

B.

The connect cluster has reached capacity, and new connectors cannot be created without expanding the cluster.

C.

The Connector already exists in the cluster.

D.

The Connect cluster is in process of rebalancing.

Buy Now
Questions 9

(You are experiencing low throughput from a Java producer.

Kafka producer metrics show a low I/O thread ratio and low I/O thread wait ratio.

What is the most likely cause of the slow producer performance?)

Options:

A.

The producer is sending large batches of messages.

B.

There is a bad data link layer (Layer 2) connection from the producer to the cluster.

C.

The producer code has an expensive callback function.

D.

Compression is enabled.

Buy Now
Questions 10

You are working on a Kafka cluster with three nodes. You create a topic named orders with:

replication.factor = 3

min.insync.replicas = 2

acks = allWhat exception will be generated if two brokers are down due to network delay?

Options:

A.

NotEnoughReplicasException

B.

NetworkException

C.

NotCoordinatorException

D.

NotLeaderForPartitionException

Buy Now
Questions 11

Which statement describes the storage location for a sink connector’s offsets?

Options:

A.

The __consumer_offsets topic, like any other consumer

B.

The topic specified in the offsets.storage.topic configuration parameter

C.

In a file specified by the offset.storage.file.filename configuration parameter

D.

In memory which is then periodically flushed to a RocksDB instance

Buy Now
Questions 12

(You have a Kafka consumer in production actively reading from a critical topic from which multiple other applications are consuming.

You have a new requirement to update the offset of your consumer to start reading from the beginning of the topic.

Which action would you take?)

Options:

A.

Update the consumer group’s offset to the earliest position using the kafka-consumer-groups CLI tool.

B.

Temporarily configure the topic’s retention.ms parameter to 0 to empty the topic.

C.

Start a new consumer application with the same consumer group id.

D.

Update the consumer configuration by setting the auto.offset.reset property to earliest.

Buy Now
Questions 13

(You create a topic with five partitions.

What can you assume about messages read from that topic by a single consumer group?)

Options:

A.

Messages can be consumed by a maximum of five consumers in the same consumer group.

B.

The consumer group can only read the same number of messages from all the partitions.

C.

All messages will be read from exactly one broker by the consumer group.

D.

Messages from one partition can be consumed by any of the consumers in a group for faster processing.

Buy Now
Questions 14

Which tool can you use to modify the replication factor of an existing topic?

Options:

A.

kafka-reassign-partitions.sh

B.

kafka-recreate-topic.sh

C.

kafka-topics.sh

D.

kafka-reassign-topics.sh

Buy Now
Questions 15

An application is consuming messages from Kafka.

The application logs show that partitions are frequently being reassigned within the consumer group.

Which two factors may be contributing to this?

(Select two.)

Options:

A.

There is a slow consumer processing application.

B.

The number of partitions does not match the number of application instances.

C.

There is a storage issue on the broker.

D.

An instance of the application is crashing and being restarted.

Buy Now
Questions 16

You have a topic t1 with six partitions. You use Kafka Connect to send data from topic t1 in your Kafka cluster to Amazon S3. Kafka Connect is configured for two tasks.

How many partitions will each task process?

Options:

A.

2

B.

3

C.

6

D.

12

Buy Now
Questions 17

Which configuration determines how many bytes of data are collected before sending messages to the Kafka broker?

Options:

A.

batch.size

B.

max.block.size

C.

buffer.memory

D.

send.buffer.bytes

Buy Now
Questions 18

Which two producer exceptions are examples of the class RetriableException? (Select two.)

Options:

A.

LeaderNotAvailableException

B.

RecordTooLargeException

C.

AuthorizationException

D.

NotEnoughReplicasException

Buy Now
Questions 19

Match the topic configuration setting with the reason the setting affects topic durability.

(You are given settings like unclean.leader.election.enable=false, replication.factor, min.insync.replicas=2)

CCDAK Question 19

Options:

Buy Now
Questions 20

(Which configuration is valid for deploying a JDBC Source Connector to read all rows from the orders table and write them to the dbl-orders topic?)

Options:

A.

{"name": "orders-connect","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl","topic.whitelist": "orders","auto.create": "true"}

B.

{"name": "dbl-orders","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.blacklist": "ord*"}

C.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&useAutoAuth=true","topic.prefix": "dbl-","table.whitelist": "orders"}

D.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.JdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.whitelist": "orders"}

Buy Now
Questions 21

(You have a Kafka Connect cluster with multiple connectors deployed.

One connector is not working as expected.

You need to find logs related to that specific connector to investigate the issue.

How can you find the connector’s logs?)

Options:

A.

Modify the log4j.properties file to enable connector context.

B.

Change the log level to DEBUG to include connector context information.

C.

Modify the log4j.properties file to add a dedicated log appender for the connector.

D.

Make no change; there is no way to isolate connector logs.

Buy Now
Questions 22

You want to connect with username and password to a secured Kafka cluster that has SSL encryption.

Which properties must your client include?

Options:

A.

security.protocol=SASL_SSLsasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

B.

security.protocol=SSLsasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

C.

security.protocol=SASL_PLAINTEXTsasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

D.

security.protocol=PLAINTEXTsasl.jaas.config=org.apache.kafka.common.security.ssl.TlsLoginModule required username='myUser' password='myPassword';

Buy Now
Questions 23

(Your configuration parameters for a Source connector and Connect worker are:

• offset.flush.interval.ms=60000

• offset.flush.timeout.ms=500

• offset.storage.topic=connect-offsets

• offset.storage.replication.factor=-1

Which two statements match the expected behavior?

Select two.)

Options:

A.

The offsets topic will use the broker default replication factor.

B.

The connector will commit offsets to the broker default offsets topic.

C.

The connector will commit offsets to a topic called connect-offsets.

D.

The connector will wait 500 ms before trying to commit offsets for tasks.

Buy Now
Questions 24

You need to explain the best reason to implement the consumer callback interface ConsumerRebalanceListener prior to a Consumer Group Rebalance.

Which statement is correct?

Options:

A.

Partitions assigned to a consumer may change.

B.

Previous log files are deleted.

C.

Offsets are compacted.

D.

Partition leaders may change.

Buy Now
Questions 25

(A stream processing application tracks user activity in online shopping carts, including items added, removed, and ordered throughout the day for each user.

You need to capture data to identify possible periods of user inactivity.

Which type of Kafka Streams window should you use?)

Options:

A.

Session

B.

Hopping

C.

Tumbling

D.

Sliding

Buy Now
Questions 26

You are creating a Kafka Streams application to process retail data.

Match the input data streams with the appropriate Kafka Streams object.

Options:

Buy Now
Questions 27

You are experiencing low throughput from a Java producer.

Metrics show low I/O thread ratio and low I/O thread wait ratio.

What is the most likely cause of the slow producer performance?

Options:

A.

Compression is enabled.

B.

The producer is sending large batches of messages.

C.

There is a bad data link layer (layer 2) connection from the producer to the cluster.

D.

The producer code has an expensive callback function.

Buy Now
Exam Code: CCDAK
Exam Name: Confluent Certified Developer for Apache Kafka Certification Examination
Last Update: Feb 20, 2026
Questions: 90

PDF + Testing Engine

$49.5  $164.99

Testing Engine

$37.5  $124.99
buy now CCDAK testing engine

PDF (Q&A)

$31.5  $104.99
buy now CCDAK pdf
dumpsmate guaranteed to pass

24/7 Customer Support

DumpsMate's team of experts is always available to respond your queries on exam preparation. Get professional answers on any topic of the certification syllabus. Our experts will thoroughly satisfy you.

Site Secure

mcafee secure

TESTED 21 Feb 2026