Download Cloudera Certified Administrator for Apache Hadoop (CCAH).CCA-500.SelfTestEngine.2020-01-15.30q.vcex

Vendor: Cloudera
Exam Code: CCA-500
Exam Name: Cloudera Certified Administrator for Apache Hadoop (CCAH)
Date: Jan 15, 2020
File Size: 211 KB
Downloads: 1

How to open VCEX files?

Files with VCEX extension can be opened by ProfExam Simulator.

Demo Questions

Question 1
You are configuring your cluster to run HDFS and MapReducer v2 (MRv2) on YARN. Which two daemons needs to be installed on your cluster’s master nodes? (Choose two)
  1. HMaster
  2. ResourceManager
  3. TaskManager
  4. JobTracker
  5. NameNode
  6. DataNode
Correct answer: BE
Question 2
You are running a Hadoop cluster with a NameNode on host mynamenode, a secondary NameNode on host mysecondarynamenode and several DataNodes. 
Which best describes how you determine when the last checkpoint happened?
  1. Execute hdfs namenode –report on the command line and look at the Last Checkpoint information
  2. Execute hdfs dfsadmin –saveNamespace on the command line which returns to you the last checkpoint value in fstime file
  3. Connect to the web UI of the Secondary NameNode (http://mysecondary:50090/) and look at the “Last Checkpoint” information
  4. Connect to the web UI of the NameNode (http://mynamenode:50070) and look at the “Last Checkpoint” information
Correct answer: C
Explanation:
Reference: https://www.inkling.com/read/hadoop-definitive-guide-tom-white-3rd/chapter-10/hdfs
Reference: https://www.inkling.com/read/hadoop-definitive-guide-tom-white-3rd/chapter-10/hdfs
Question 3
What does CDH packaging do on install to facilitate Kerberos security setup?
  1. Automatically configures permissions for log files at & MAPRED_LOG_DIR/userlogs
  2. Creates users for hdfs and mapreduce to facilitate role assignment
  3. Creates directories for temp, hdfs, and mapreduce with the correct permissions
  4. Creates a set of pre-configured Kerberos keytab files and their permissions
  5. Creates and configures your kdc with default cluster values
Correct answer: B
Question 4
Which three basic configuration parameters must you set to migrate your cluster from MapReduce 1 (MRv1) to MapReduce V2 (MRv2)? (Choose three)
  1. Configure the NodeManager to enable MapReduce services on YARN by setting the following property in yarn-site.xml: 
    <name>yarn.nodemanager.hostname</name> 
    <value>your_nodeManager_shuffle</value>
  2. Configure the NodeManager hostname and enable node services on YARN by setting the following property in yarn-site.xml: 
    <name>yarn.nodemanager.hostname</name> 
    <value>your_nodeManager_hostname</value>
  3. Configure a default scheduler to run on YARN by setting the following property in mapred-site.xml: 
    <name>mapreduce.jobtracker.taskScheduler</name> 
    <Value>org.apache.hadoop.mapred.JobQueueTaskScheduler</value>
  4. Configure the number of map tasks per jon YARN by setting the following property in mapred: 
    <name>mapreduce.job.maps</name> 
    <value>2</value>
  5. Configure the ResourceManager hostname and enable node services on YARN by setting the following property in yarn-site.xml: 
    <name>yarn.resourcemanager.hostname</name> 
    <value>your_resourceManager_hostname</value>
  6. Configure MapReduce as a Framework running on YARN by setting the following property in mapred-site.xml: 
    <name>mapreduce.framework.name</name> 
    <value>yarn</value>
Correct answer: AEF
Question 5
You need to analyze 60,000,000 images stored in JPEG format, each of which is approximately 25 KB. Because you Hadoop cluster isn’t optimized for storing and processing many small files, you decide to do the following actions: 
  1. Group the individual images into a set of larger files 
  2. Use the set of larger files as input for a MapReduce job that processes them directly with python using Hadoop streaming. 
Which data serialization system gives the flexibility to do this?
  1. CSV
  2. XML
  3. HTML
  4. Avro
  5. SequenceFiles
  6. JSON
Correct answer: E
Explanation:
Sequence files are block-compressed and provide direct serialization and deserialization of several arbitrary data types (not just text). Sequence files can be generated as the output of other MapReduce tasks and are an efficient intermediate representation for data that is passing from one MapReduce job to anther.
Sequence files are block-compressed and provide direct serialization and deserialization of several arbitrary data types (not just text). Sequence files can be generated as the output of other MapReduce tasks and are an efficient intermediate representation for data that is passing from one MapReduce job to anther.
Question 6
Identify two features/issues that YARN is designated to address: (Choose two)
  1. Standardize on a single MapReduce API
  2. Single point of failure in the NameNode
  3. Reduce complexity of the MapReduce APIs
  4. Resource pressure on the JobTracker
  5. Ability to run framework other than MapReduce, such as MPI
  6. HDFS latency
Correct answer: DE
Explanation:
Reference: http://www.revelytix.com/?q=content/hadoop-ecosystem (YARN, first para)
Reference: http://www.revelytix.com/?q=content/hadoop-ecosystem (YARN, first para)
Question 7
Which is the default scheduler in YARN?
  1. YARN doesn’t configure a default scheduler, you must first assign an appropriate scheduler class in yarn-site.xml
  2. Capacity Scheduler
  3. Fair Scheduler
  4. FIFO Scheduler
Correct answer: B
Explanation:
Reference: http://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-site/CapacityScheduler.html
Reference: http://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-site/CapacityScheduler.html
Question 8
Which scheduler would you deploy to ensure that your cluster allows short jobs to finish within a reasonable time without starting long-running jobs?
  1. Complexity Fair Scheduler (CFS)
  2. Capacity Scheduler
  3. Fair Scheduler
  4. FIFO Scheduler
Correct answer: C
Explanation:
Reference: http://hadoop.apache.org/docs/r1.2.1/fair_scheduler.html
Reference: http://hadoop.apache.org/docs/r1.2.1/fair_scheduler.html
Question 9
Your cluster is configured with HDFS and MapReduce version 2 (MRv2) on YARN. What is the result when you execute: hadoop jar SampleJar MyClass on a client machine?
  1. SampleJar.Jar is sent to the ApplicationMaster which allocates a container for SampleJar.Jar
  2. Sample.jar is placed in a temporary directory in HDFS
  3. SampleJar.jar is sent directly to the ResourceManager
  4. SampleJar.jar is serialized into an XML file which is submitted to the ApplicatoionMaster
Correct answer: A
Question 10
You are working on a project where you need to chain together MapReduce, Pig jobs. You also need the ability to use forks, decision points, and path joins. Which ecosystem project should you use to perform these actions?
  1. Oozie
  2. ZooKeeper
  3. HBase
  4. Sqoop
  5. HUE
Correct answer: A
HOW TO OPEN VCE FILES

Use VCE Exam Simulator to open VCE files
Avanaset

HOW TO OPEN VCEX AND EXAM FILES

Use ProfExam Simulator to open VCEX and EXAM files
ProfExam Screen

ProfExam
ProfExam at a 20% markdown

You have the opportunity to purchase ProfExam at a 20% reduced price

Get Now!