Latest 156-315.81 Practice Tests with Actual Questions

Get Complete pool of questions with Premium PDF and Test Engine

Exam Code : 156-315.81
Exam Name : Check Point Certified Security Expert R81
Vendor Name : "Checkpoint"







156-315.81 Dumps

156-315.81 Braindumps

156-315.81 Real Questions

156-315.81 Practice Test

156-315.81 Actual Questions


killexams.com


Checkpoint


156-315.81


Check Point Certified Security Expert R81


https://killexams.com/pass4sure/exam-detail/156-315.81



Question: 53


A Zookeeper ensemble contains 3 servers.


Over which ports the members of the ensemble should be able to communicate in default configuration? (select three) A. 2181

  1. 3888

  2. 443

  3. 2888

  4. 9092

  5. 80




Answer: A,B,D
Explanation:

2181 – client port, 2888 – peer port, 3888 – leader port



Question: 54


You are running a Kafka Streams application in a Docker container managed by Kubernetes, and upon application restart, it takes a long time for the docker container to replicate the state and get back to processing the data .


How can you improve dramatically the application restart?

  1. Mount a persistent volume for your RocksDB

  2. Increase the number of partitions in your inputs topic

  3. Reduce the Streams caching property

  4. Increase the number of Streams threads




Answer: A
Explanation:

Although any Kafka Streams application is stateless as the state is stored in Kafka, it can take a while and lots of resources to recover the state from Kafka. In order to speed up recovery, it is advised to store the Kafka Streams state on a persistent volume, so that only the missing part of the state needs to be recovered.



Question: 55


A producer is sending messages with null key to a topic with 6 partitions using the DefaultPartitioner. Where will the messages be stored?

  1. Partition 5

  2. Any of the topic partitions

  3. The partition for the null key

  4. Partition 0




Answer: A
Explanation:

Message with no keys will be stored with round-robin strategy among partitions.



Question: 56

A kafka topic has a replication factor of 3 and min.insync.replicas setting of 2 . How many brokers can go down before a producer with acks=all can’t produce?

  1. 0

  2. 2

  3. 1

  4. 3




Answer: C
Explanation:

acks=all and min.insync.replicas=2 means we must have at least 2 brokers up for the partition to be available



Question: 57


How can you gracefully make a Kafka consumer to stop immediately polling data from Kafka and gracefully shut down a consumer application?

  1. Call consumer.wakeUp() and catch a WakeUpException

  2. Call consumer.poll() in another thread

  3. Kill the consumer thread




Answer: A
Explanation:

See https://stackoverflow.com/a/37748336/3019499



Question: 58


A Zookeeper ensemble contains 5 servers .


What is the maximum number of servers that can go missing and the ensemble still run?

  1. 3

  2. 4

  3. 2

  4. 1




Answer: C
Explanation:

majority consists of 3 zk nodes for 5 nodes zk cluster, so 2 can fail



Question: 59


You have a Kafka cluster and all the topics have a replication factor of 3. One intern at your company stopped a broker, and accidentally deleted all the data of that broker on the disk .


What will happen if the broker is restarted?

  1. The broker will start, and other topics will also be deleted as the broker data on the disk got deleted

  2. The broker will start, and won’t be online until all the data it needs to have is replicated from other leaders

  3. The broker will crash

  4. The broker will start, and won’t have any data. If the broker comes leader, we have a data loss




Answer: B
Explanation:

Kafka replication mechanism makes it resilient to the scenarios where the broker lose data on disk, but can recover from replicating from other brokers. This makes Kafka amazing!



Question: 60

In the Kafka consumer metrics it is observed that fetch-rate is very high and each fetch is small . What steps will you take to increase throughput?

  1. Increase fetch.max.wait

  2. Increase fetch.max.bytes

  3. Decrease fetch.max.bytes

  4. Decrease fetch.min.bytes

  5. Increase fetch.min.bytes




Answer: E
Explanation:

This will allow consumers to wait and receive more bytes in each fetch request.