Get all latest (August) Cloudera CCA-500 Actual Test 11-20

Ensurepass

 

QUESTION 11

You have A 20 node Hadoop cluster, with 18 slave nodes and 2 master nodes running HDFS High Availability (HA). You want to minimize the chance of data loss in your cluster. What should you do?

 

A.

Add another master node to increase the number of nodes running the JournalNode which increases the number of machines available to HA to create a quorum.

B.

Set an HDFS replication factor that provides data redundancy, protecting against node failure.

C.

Run a Secondary NameNode on a different master from the NameNode in order to provide automatic recovery from a NameNode failure.

D.

Run the ResourceManager on a different master from the NameNode in order to load- share HDFS metadata processing.

E.

Configure the cluster’s disk drives with an appropriate fault tolerant RAID level.

 

Correct Answer: D

 

 

QUESTION 12

On a cluster running MapReduce v2 (MRv2) on YARN, a MapReduce job is given a directory of 10 plain text files as its input directory. Each file is made up of 3 HDFS blocks. How many Mappers will run?

 

A.

We cannot say; the number of Mappers is determined by the ResourceManager

B.

We cannot say; the number of Mappers is determined by the developer

C.

30

D.

3

E.

10

F.

We cannot say; the number of mappers is determined by the ApplicationMaster

 

Correct Answer: E

QUESTION 13

Your Hadoop cluster contains nodes in three racks. You have not configured the dfs.hosts property in the NameNode’s configuration file. What results?

 

A.

The NameNode will update the dfs.hosts property to include machines running the DataNode daemon on the next NameNode reboot or with the command dfsadmin -refreshNodes

B.

No new nodes can be added to the cluster until you specify them in the dfs.hosts file

C.

Any machine running the DataNode daemon can immediately join the cluster

D.

Presented with a blank dfs.hosts property, the NameNode will permit DataNodes specified in mapred.hosts to join the cluster

 

Correct Answer: C

 

 

QUESTION 14

You are configuring your cluster to run HDFS and MapReducer v2 (MRv2) on YARN. Which two daemons needs to be installed on your cluster’s master nodes? (Choose two)

 

A.

HMaster

B.

ResourceManager

C.

TaskManager

D.

JobTracker

E.

NameNode

F.

DataNode

 

Correct Answer: BE

 

 

QUESTION 15

Which two features does Kerberos security add to a Hadoop cluster? (Choose two)

 

A.

User authentication on all remote procedure calls (RPCs)

B.

Encryption for data during transfer between the Mappers and Reducers

C.

Encryption for data on disk (“at rest”)

D.

Authentication for user access to the cluster against a central server

E.

Root access to the cluster for users hdfs and mapred but non-root access for clients

 

Correct Answer: AD

 

 

QUESTION 16

Which YARN daemon or service monitors a Controller’s per-application resource using (e.g., memory CPU)?

 

A.

ApplicationMaster

B.

NodeManager

C.

ApplicationManagerService

D.

ResourceManager

 

Correct Answer: A

 

 

 

QUESTION 17

You want to understand more about how users browse your public website. For example, you want to know which pages they visit prior to placing an order. You have a server farm of 200 web servers hosting your website. Which is the most efficient process to gather these web server across logs into your Hadoop cluster analysis?

 

A.

Sample the web server logs web servers and copy them into HDFS using curl

B.

Ingest the server web logs into HDFS using Flume

C.

Channel these clickstreams into Hadoop using Hadoop Streaming

D.

Import all user clicks from your OLTP databases into Hadoop using Sqoop

E.

Write a MapReeeduce job with the web servers for mappers and the Hadoop cluster nodes for reducers

 

Correct Answer: B

Explanation:

Apache Flume is a service for streaming logs into Hadoop. Apache Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of streaming data into the Hadoop Distributed File System (HDFS). It has a simple and flexible architecture based on streaming data flows; and is robust and fault tolerant with tunable reliability mechanisms for failover and recovery.

 

 

QUESTION 18

Cluster Summary:

 

45 files and directories, 12 blocks = 57 total. Heap size is 15.31 MB/193.38MB(7%)

 

Refer to the above screenshot.

 

You configure a Hadoop cluster with seven DataNodes and on of your monitoring UIs displays the details shown in the exhibit.

 

What does the this tell you?

 

clip_image002

 

A.

The DataNode JVM on one host is not active.

B.

Because your under-replicated blocks count matches the Live Nodes, one node is dead, and your DFS Used % equals 0%, you can’t be certain that your cluster has all the data you’ve written it.

C.

Your cluster has lost all HDFS data which had bocks stored on the dead DatNode.

D.

The HDFS cluster is in safe mode.

Correct Answer: A

 

 

QUESTION 19

Identify two features/issues that YARN is designated to address: (Choose two)

 

A.

Standardize on a single MapReduce API

B.

Single point of failure in the NameNode

C.

Reduce complexity of the MapReduce APIs

D.

Resource pressure on the JobTracker

E.

Ability to run framework other than MapReduce, such as MPI

F.

HDFS latency

 

Correct Answer: DE

Explanation:

http://www.revelytix.com/?q=content/hadoop-ecosystem(YARN, first para)

 

 

QUESTION 20

Assume you have a file named foo.txt in your local directory. You issue the following three commands:

 

Hadoop fs -mkdir input

 

Hadoop fs -put foo.txt input/foo.txt

 

Hadoop fs -put foo.txt input

 

What happens when you issue the third command?

 

A.

The write succeeds, overwriting foo.txt in HDFS with no warning

B.

The file is uploaded and stored as a plain file named input

C.

You get a warning that foo.txt is being overwritten

D.

You get an error message telling you that foo.txt already exists, and asking you if you would like to overwrite it.

E.

You get a error message telling you that foo.txt already exists. The file is not written to HDFS

F.

You get an error message telling you that input is not a directory

G.

The write silently fails

 

Correct Answer: CE

 

Free VCE & PDF File for Cloudera CCA-500 Real Exam

Instant Access to Free VCE Files: CompTIA | VMware | SAP …
Instant Access to Free PDF Files: CompTIA | VMware | SAP …

This entry was posted in CCA-500 Real Exam (August) and tagged , , , , , , . Bookmark the permalink.