Cloudera Certified Administrator For Apache Hadoop (CCAH) CCA-500 Exam Questions And Answers

Want to know features? Want to lear more about experience? Study . Gat a success with an absolute guarantee to pass Cloudera CCA-500 (Cloudera Certified Administrator for Apache Hadoop (CCAH)) test on your first attempt.

Free CCA-500 Demo Online For Microsoft Certifitcation:

Page: 1 / 5
Total 60 questions Full Exam Access
Question 1
You are running Hadoop cluster with all monitoring facilities properly configured. Which scenario will go undeselected?
My answer: -
Reference answer: B
Reference analysis:

None

Question 2
You are configuring a server running HDFS, MapReduce version 2 (MRv2) on YARN running Linux. How must you format underlying file system of each DataNode?
My answer: -
Reference answer: B
Reference analysis:

None

Question 3
You have a Hadoop cluster HDFS, and a gateway machine external to the cluster from which clients submit jobs. What do you need to do in order to run Impala on the cluster and submit jobs from the command line of the gateway machine?
My answer: -
Reference answer: D
Reference analysis:

None

Question 4
Your cluster’s mapred-start.xml includes the following parameters
My answer: -
Reference answer: D
Reference analysis:

None

Question 5
You want to understand more about how users browse your public website. For example, you want to know which pages they visit prior to placing an order. You have a server farm of 200 web servers hosting your website. Which is the most efficient process to gather these web server across logs into your Hadoop cluster analysis?
My answer: -
Reference answer: B
Reference analysis:

Apache Flume is a service for streaming logs into Hadoop.
Apache Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of streaming data into the Hadoop Distributed File System (HDFS). It has a simple and flexible architecture based on streaming data flows; and is robust and fault tolerant with tunable reliability mechanisms for failover and recovery.

Question 6
On a cluster running MapReduce v2 (MRv2) on YARN, a MapReduce job is given a directory of 10 plain text files as its input directory. Each file is made up of 3 HDFS blocks. How many Mappers will run?
My answer: -
Reference answer: E
Reference analysis:

None

Question 7
You have A 20 node Hadoop cluster, with 18 slave nodes and 2 master nodes running HDFS High Availability (HA). You want to minimize the chance of data loss in your cluster. What should you do?
My answer: -
Reference answer: D
Reference analysis:

None

Question 8
Which two features does Kerberos security add to a Hadoop cluster?(Choose two)
My answer: -
Reference answer: AD
Reference analysis:

None

Question 9
You have installed a cluster HDFS and MapReduce version 2 (MRv2) on YARN. You have no dfs.hosts entry(ies) in your hdfs-site.xml configuration file. You configure a new worker node by setting fs.default.name in its configuration files to point to the NameNode on your cluster, and you start the DataNode daemon on that worker node. What do you have to do on the cluster to allow the worker node to join, and start sorting HDFS blocks?
My answer: -
Reference answer: A
Reference analysis:

None

Question 10
You’re upgrading a Hadoop cluster from HDFS and MapReduce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce a block size of 128MB for all new files written to the cluster after upgrade. What should you do?
My answer: -
Reference answer: C
Reference analysis:

None

Question 11
On a cluster running CDH 5.0 or above, you use the hadoop fs –put command to write a 300MB file into a previously empty directory using an HDFS block size of 64 MB. Just after this command has finished writing 200 MB of this file, what would another use see when they look in directory?
My answer: -
Reference answer: B
Reference analysis:

None

Question 12
You decide to create a cluster which runs HDFS in High Availability mode with automatic failover, using Quorum Storage. What is the purpose of ZooKeeper in such a configuration?
My answer: -
Reference answer: A
Reference analysis:

Reference: Reference:http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/PDF/CDH4-High-Availability-Guide.pdf(page 15)

Question 13
Table schemas in Hive are:
My answer: -
Reference answer: B
Reference analysis:

None

Question 14
You want to node to only swap Hadoop daemon data from RAM to disk when absolutely necessary. What should you do?
My answer: -
Reference answer: D
Reference analysis:

None

Question 15
Given:
CCA-500 dumps exhibit
You want to clean up this list by removing jobs where the State is KILLED. What command you enter?
My answer: -
Reference answer: B
Reference analysis:

Reference:http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1-latest/bk_using-apache-hadoop/content/common_mrv2_commands.html

Question 16
A slave node in your cluster has 4 TB hard drives installed (4 x 2TB). The DataNode is configured to store HDFS blocks on all disks. You set the value of the dfs.datanode.du.reserved parameter to 100 GB. How does this alter HDFS block storage?
My answer: -
Reference answer: B
Reference analysis:

None

Question 17
You have just run a MapReduce job to filter user messages to only those of a selected geographical region. The output for this job is in a directory named westUsers, located just below your home directory in HDFS. Which command gathers these into a single file on your local file system?
My answer: -
Reference answer: B
Reference analysis:

None

Question 18
Which three basic configuration parameters must you set to migrate your cluster from MapReduce 1 (MRv1) to MapReduce V2 (MRv2)?(Choose three)
My answer: -
Reference answer: AEF
Reference analysis:

None

Question 19
Each node in your Hadoop cluster, running YARN, has 64GB memory and 24 cores. Your yarn.site.xml has the following configuration:
My answer: -
Reference answer: A
Reference analysis:

None

Question 20
Your Hadoop cluster is configuring with HDFS and MapReduce version 2 (MRv2) on YARN. Can you configure a worker node to run a NodeManager daemon but not a DataNode daemon and still have a functional cluster?
My answer: -
Reference answer: B
Reference analysis:

None

Question 21
For each YARN job, the Hadoop framework generates task log file. Where are Hadoop task log files stored?
My answer: -
Reference answer: D
Reference analysis:

None

Question 22
You are planning a Hadoop cluster and considering implementing 10 Gigabit Ethernet as the network fabric. Which workloads benefit the most from faster network fabric?
My answer: -
Reference answer: A
Reference analysis:

None

Page: 1 / 5
Total 60 questions Full Exam Access