Summer Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Free and Premium Confluent CCDAK Dumps Questions Answers

Page: 1 / 5
Total 61 questions

Confluent Certified Developer for Apache Kafka Certification Examination Questions and Answers

Question 1

Which partition assignment minimizes partition movements between two assignments?

Options:

A.

RoundRobinAssignor

B.

StickyAssignor

C.

RangeAssignor

D.

PartitionAssignor

Buy Now
Question 2

Which statement is true about how exactly-once semantics (EOS) work in Kafka Streams?

Options:

A.

Kafka Streams disables log compaction on internal changelog topics to preserve all state changes for potential recovery.

B.

EOS in Kafka Streams relies on transactional producers to atomically commit state updates to changelog topics and output records to Kafka.

C.

Kafka Streams provides EOS by periodically checkpointing state stores and replaying changelogs to recover only unprocessed messages during failure.

D.

EOS in Kafka Streams is implemented by creating a separate Kafka topic for deduplication of all messages processed by the application.

Question 3

Clients that connect to a Kafka cluster are required to specify one or more brokers in the bootstrap.servers parameter.

What is the primary advantage of specifying more than one broker?

Options:

A.

It provides redundancy in making the initial connection to the Kafka cluster.

B.

It forces clients to enumerate every single broker in the cluster.

C.

It is the mechanism to distribute a topic’s partitions across multiple brokers.

D.

It provides the ability to wake up dormant brokers.

Question 4

Which two statements about Kafka Connect Single Message Transforms (SMTs) are correct?

(Select two.)

Options:

A.

Multiple SMTs can be chained together and act on source or sink messages.

B.

SMTs are often used to join multiple records from a source data system into a single Kafka record.

C.

Masking data is a good example of an SMT.

D.

SMT functionality is included within Kafka Connect converters.

Question 5

You need to explain the best reason to implement the consumer callback interface ConsumerRebalanceListener prior to a Consumer Group Rebalance.

Which statement is correct?

Options:

A.

Partitions assigned to a consumer may change.

B.

Previous log files are deleted.

C.

Offsets are compacted.

D.

Partition leaders may change.

Question 6

You need to set alerts on key broker metrics to trigger notifications when the cluster is unhealthy.

Which are three minimum broker metrics to monitor?

(Select three.)

Options:

A.

kafka.controller:type=KafkaController,name=TopicsToDeleteCount

B.

kafka.controller:type=KafkaController,name=OfflinePartitionsCount

C.

kafka.controller:type=KafkaController,name=ActiveControllerCount

D.

kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec

E.

kafka.controller:type=KafkaController,name=LastCommittedRecordOffset

Question 7

A producer is configured with the default partitioner. It is sending records to a topic that is configured with five partitions. The record does not contain any key.

What is the result of this?

Options:

A.

Records will be dispatched among the available partitions.

B.

Records will be sent to partition 0.

C.

An error will be raised and no record will be sent.

D.

Records will be sent to the least used partition.

Question 8

Match the topic configuration setting with the reason the setting affects topic durability.

(You are given settings like unclean.leader.election.enable=false, replication.factor, min.insync.replicas=2)

Options:

Question 9

Which two producer exceptions are examples of the class RetriableException? (Select two.)

Options:

A.

LeaderNotAvailableException

B.

RecordTooLargeException

C.

AuthorizationException

D.

NotEnoughReplicasException

Question 10

Which two statements are correct when assigning partitions to the consumers in a consumer group using the assign() API?

(Select two.)

Options:

A.

It is mandatory to subscribe to a topic before calling assign() to assign partitions.

B.

The consumer chooses which partition to read without any assignment from brokers.

C.

The consumer group will not be rebalanced if a consumer leaves the group.

D.

All topics must have the same number of partitions to use assign() API.

Question 11

What is the default maximum size of a message the Apache Kafka broker can accept?

Options:

A.

1MB

B.

2MB

C.

5MB

D.

10MB

Question 12

Where are source connector offsets stored?

Options:

A.

offset.storage.topic

B.

storage.offset.topic

C.

topic.offset.config

D.

offset, storage, partitions

Question 13

A stream processing application is consuming from a topic with five partitions. You run three instances of the application. Each instance has num.stream.threads=5.

You need to identify the number of stream tasks that will be created and how many will actively consume messages from the input topic.

Options:

A.

5 created, 1 actively consuming

B.

5 created, 5 actively consuming

C.

15 created, 5 actively consuming

D.

15 created, 15 actively consuming

Question 14

Which statement describes the storage location for a sink connector’s offsets?

Options:

A.

The __consumer_offsets topic, like any other consumer

B.

The topic specified in the offsets.storage.topic configuration parameter

C.

In a file specified by the offset.storage.file.filename configuration parameter

D.

In memory which is then periodically flushed to a RocksDB instance

Question 15

Which two producer exceptions are examples of the class RetriableException? (Select two.)

Options:

A.

LeaderNotAvailableException

B.

RecordTooLargeException

C.

AuthorizationException

D.

NotEnoughReplicasException

Question 16

This schema excerpt is an example of which schema format?

package com.mycorp.mynamespace;

message SampleRecord {

int32 Stock = 1;

double Price = 2;

string Product_Name = 3;

}

Options:

A.

Avro

B.

Protobuf

C.

JSON Schema

D.

YAML

Question 17

You are writing a producer application and need to ensure proper delivery. You configure the producer with acks=all.

Which two actions should you take to ensure proper error handling?

(Select two.)

Options:

A.

Use a callback argument in producer.send() where you check delivery status.

B.

Check that producer.send() returned a RecordMetadata object and is not null.

C.

Surround the call of producer.send() with a try/catch block to catch KafkaException.

D.

Check the value of ProducerRecord.status().

Question 18

What is a consequence of increasing the number of partitions in an existing Kafka topic?

Options:

A.

Existing data will be redistributed across the new number of partitions temporarily increasing cluster load.

B.

Records with the same key could be located in different partitions.

C.

Consumers will need to process data from more partitions which will significantly increase consumer lag.

D.

The acknowledgment process will increase latency for producers using acks=all.

Page: 1 / 5
Total 61 questions