Independence Day Special 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: save70

Cloudera CCA-500 Dumps

Page: 1 / 2
Total 60 questions

Cloudera Certified Administrator for Apache Hadoop (CCAH) Questions and Answers

Question 1

Which command does Hadoop offer to discover missing or corrupt HDFS data?

Options:

A.

Hdfs fs –du

B.

Hdfs fsck

C.

Dskchk

D.

The map-only checksum

E.

Hadoop does not provide any tools to discover missing or corrupt data; there is not need because three replicas are kept for each data block

Question 2

What does CDH packaging do on install to facilitate Kerberos security setup?

Options:

A.

Automatically configures permissions for log files at & MAPRED_LOG_DIR/userlogs

B.

Creates users for hdfs and mapreduce to facilitate role assignment

C.

Creates directories for temp, hdfs, and mapreduce with the correct permissions

D.

Creates a set of pre-configured Kerberos keytab files and their permissions

E.

Creates and configures your kdc with default cluster values

Question 3

During the execution of a MapReduce v2 (MRv2) job on YARN, where does the Mapper place the intermediate data of each Map Task?

Options:

A.

The Mapper stores the intermediate data on the node running the Job’s ApplicationMaster so that it is available to YARN ShuffleService before the data is presented to the Reducer

B.

The Mapper stores the intermediate data in HDFS on the node where the Map tasks ran in the HDFS /usercache/&(user)/apache/application_&(appid) directory for the user who ran the job

C.

The Mapper transfers the intermediate data immediately to the reducers as it is generated by the Map Task

D.

YARN holds the intermediate data in the NodeManager’s memory (a container) until it is transferred to the Reducer

E.

The Mapper stores the intermediate data on the underlying filesystem of the local disk in the directories yarn.nodemanager.locak-DIFS

Question 4

You are running a Hadoop cluster with a NameNode on host mynamenode. What are two ways to determine available HDFS space in your cluster?

Options:

A.

Run hdfs fs –du / and locate the DFS Remaining value

B.

Run hdfs dfsadmin –report and locate the DFS Remaining value

C.

Run hdfs dfs / and subtract NDFS Used from configured Capacity

D.

Connect to http://mynamenode:50070/dfshealth.jsp and locate the DFS remaining value

Question 5

Which YARN daemon or service monitors a Controller’s per-application resource using (e.g., memory CPU)?

Options:

A.

ApplicationMaster

B.

NodeManager

C.

ApplicationManagerService

D.

ResourceManager

Question 6

You are planning a Hadoop cluster and considering implementing 10 Gigabit Ethernet as the network fabric. Which workloads benefit the most from faster network fabric?

Options:

A.

When your workload generates a large amount of output data, significantly larger than the amount of intermediate data

B.

When your workload consumes a large amount of input data, relative to the entire capacity if HDFS

C.

When your workload consists of processor-intensive tasks

D.

When your workload generates a large amount of intermediate data, on the order of the input data itself

Question 7

For each YARN job, the Hadoop framework generates task log file. Where are Hadoop task log files stored?

Options:

A.

Cached by the NodeManager managing the job containers, then written to a log directory on the NameNode

B.

Cached in the YARN container running the task, then copied into HDFS on job completion

C.

In HDFS, in the directory of the user who generates the job

D.

On the local disk of the slave mode running the task

Question 8

You have a cluster running with a FIFO scheduler enabled. You submit a large job A to the cluster, which you expect to run for one hour. Then, you submit job B to the cluster, which you expect to run a couple of minutes only.

You submit both jobs with the same priority.

Which two best describes how FIFO Scheduler arbitrates the cluster resources for job and its tasks? (Choose two)

Options:

A.

Because there is a more than a single job on the cluster, the FIFO Scheduler will enforce a limit on the percentage of resources allocated to a particular job at any given time

B.

Tasks are scheduled on the order of their job submission

C.

The order of execution of job may vary

D.

Given job A and submitted in that order, all tasks from job A are guaranteed to finish before all tasks from job B

E.

The FIFO Scheduler will give, on average, and equal share of the cluster resources over the job lifecycle

F.

The FIFO Scheduler will pass an exception back to the client when Job B is submitted, since all slots on the cluster are use

Question 9

You want to node to only swap Hadoop daemon data from RAM to disk when absolutely necessary. What should you do?

Options:

A.

Delete the /dev/vmswap file on the node

B.

Delete the /etc/swap file on the node

C.

Set the ram.swap parameter to 0 in core-site.xml

D.

Set vm.swapfile file on the node

E.

Delete the /swapfile file on the node

Page: 1 / 2
Total 60 questions