Weekend Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code = simple70

Pass the Cloudera CCAH CCA-500 Questions and answers with ExamsMirror

Practice at least 50% of the questions to maximize your chances of passing.
Exam CCA-500 Premium Access

View all detail and faqs for the CCA-500 exam


390 Students Passed

90% Average Score

90% Same Questions
Viewing page 1 out of 2 pages
Viewing questions 1-10 out of questions
Questions # 1:

Which scheduler would you deploy to ensure that your cluster allows short jobs to finish within a reasonable time without starting long-running jobs?

Options:

A.

Complexity Fair Scheduler (CFS)

B.

Capacity Scheduler

C.

Fair Scheduler

D.

FIFO Scheduler

Questions # 2:

You have just run a MapReduce job to filter user messages to only those of a selected geographical region. The output for this job is in a directory named westUsers, located just below your home directory in HDFS. Which command gathers these into a single file on your local file system?

Options:

A.

Hadoop fs –getmerge –R westUsers.txt

B.

Hadoop fs –getemerge westUsers westUsers.txt

C.

Hadoop fs –cp westUsers/* westUsers.txt

D.

Hadoop fs –get westUsers westUsers.txt

Questions # 3:

You are working on a project where you need to chain together MapReduce, Pig jobs. You also need the ability to use forks, decision points, and path joins. Which ecosystem project should you use to perform these actions?

Options:

A.

Oozie

B.

ZooKeeper

C.

HBase

D.

Sqoop

E.

HUE

Questions # 4:

You are running a Hadoop cluster with a NameNode on host mynamenode, a secondary NameNode on host mysecondarynamenode and several DataNodes.

Which best describes how you determine when the last checkpoint happened?

Options:

A.

Execute hdfs namenode –report on the command line and look at the Last Checkpoint information

B.

Execute hdfs dfsadmin –saveNamespace on the command line which returns to you the last checkpoint value in fstime file

C.

Connect to the web UI of the Secondary NameNode (http://mysecondary:50090/) and look at the “Last Checkpoint” information

D.

Connect to the web UI of the NameNode (http://mynamenode:50070) and look at the “Last Checkpoint” information

Questions # 5:

Your Hadoop cluster contains nodes in three racks. You have not configured the dfs.hosts property in the NameNode’s configuration file. What results?

Options:

A.

The NameNode will update the dfs.hosts property to include machines running the DataNode daemon on the next NameNode reboot or with the command dfsadmin –refreshNodes

B.

No new nodes can be added to the cluster until you specify them in the dfs.hosts file

C.

Any machine running the DataNode daemon can immediately join the cluster

D.

Presented with a blank dfs.hosts property, the NameNode will permit DataNodes specified in mapred.hosts to join the cluster

Questions # 6:

Your cluster implements HDFS High Availability (HA). Your two NameNodes are named nn01 and nn02. What occurs when you execute the command: hdfs haadmin –failover nn01 nn02?

Options:

A.

nn02 is fenced, and nn01 becomes the active NameNode

B.

nn01 is fenced, and nn02 becomes the active NameNode

C.

nn01 becomes the standby NameNode and nn02 becomes the active NameNode

D.

nn02 becomes the standby NameNode and nn01 becomes the active NameNode

Questions # 7:

You suspect that your NameNode is incorrectly configured, and is swapping memory to disk. Which Linux commands help you to identify whether swapping is occurring? (Select all that apply)

Options:

A.

free

B.

df

C.

memcat

D.

top

E.

jps

F.

vmstat

G.

swapinfo

Questions # 8:

You are planning a Hadoop cluster and considering implementing 10 Gigabit Ethernet as the network fabric. Which workloads benefit the most from faster network fabric?

Options:

A.

When your workload generates a large amount of output data, significantly larger than the amount of intermediate data

B.

When your workload consumes a large amount of input data, relative to the entire capacity if HDFS

C.

When your workload consists of processor-intensive tasks

D.

When your workload generates a large amount of intermediate data, on the order of the input data itself

Questions # 9:

Which two are features of Hadoop’s rack topology? (Choose two)

Options:

A.

Configuration of rack awareness is accomplished using a configuration file. You cannot use a rack topology script.

B.

Hadoop gives preference to intra-rack data transfer in order to conserve bandwidth

C.

Rack location is considered in the HDFS block placement policy

D.

HDFS is rack aware but MapReduce daemon are not

E.

Even for small clusters on a single rack, configuring rack awareness will improve performance

Questions # 10:

You have a cluster running with a FIFO scheduler enabled. You submit a large job A to the cluster, which you expect to run for one hour. Then, you submit job B to the cluster, which you expect to run a couple of minutes only.

You submit both jobs with the same priority.

Which two best describes how FIFO Scheduler arbitrates the cluster resources for job and its tasks? (Choose two)

Options:

A.

Because there is a more than a single job on the cluster, the FIFO Scheduler will enforce a limit on the percentage of resources allocated to a particular job at any given time

B.

Tasks are scheduled on the order of their job submission

C.

The order of execution of job may vary

D.

Given job A and submitted in that order, all tasks from job A are guaranteed to finish before all tasks from job B

E.

The FIFO Scheduler will give, on average, and equal share of the cluster resources over the job lifecycle

F.

The FIFO Scheduler will pass an exception back to the client when Job B is submitted, since all slots on the cluster are use

Viewing page 1 out of 2 pages
Viewing questions 1-10 out of questions
TOP CODES

TOP CODES

Top selling exam codes in the certification world, popular, in demand and updated to help you pass on the first try.