What is the cause of the error?
A user comes to you, complaining that when she attempts to submit a Hadoop job, it fails.
There is a Directory in HDFS named /data/input. The Jar is named j.jar, and the driver class
is named DriverClass. She runs the command: Hadoop jar j.jar DriverClass
/data/input/data/output The error message returned includes the line:
PriviligedActionException as:training (auth:SIMPLE)
cause:org.apache.hadoop.mapreduce.lib.input.invalidInputException: Input path does not
exist: file:/data/input What is the cause of the error?
What is the best way to obtain and ingest these user records?
Your company stores user profile records in an OLTP databases. You want to join these
records with web server logs you have already ingested into the Hadoop file system. What
is the best way to obtain and ingest these user records?
Which two are features of Hadoop’s rack topology?
Which two are features of Hadoop’s rack topology?
Identify the function performed by the Secondary NameNode daemon on a cluster configured to run with a single
Identify the function performed by the Secondary NameNode daemon on a cluster configured to
run with a single NameNode.
Which MapReduce daemon instantiates user code, and executes map and reduce tasks on a cluster running MapReduc
Which MapReduce daemon instantiates user code, and executes map and reduce tasks on a
cluster running MapReduce v1 (MRv1)?
which best describe a Hadoop cluster’s block size storage parameters once you set the HDFS default block
Choose which best describe a Hadoop cluster’s block size storage parameters once you set the
HDFS default block size to 64MB?
How does the data get from the DataNodes to the client?
A client wants to read a file from HDFS. How does the data get from the DataNodes to the client?
which scenario should not result in loss of HDFS data assuming the default replication factor settings?
Assuming a large properly configured multi-rack Hadoop cluster, which scenario should not result
in loss of HDFS data assuming the default replication factor settings?
What is the recommended disk configuration for slave nodes in your Hadoop cluster with 6 x 2 TB hard drives?
What is the recommended disk configuration for slave nodes in your Hadoop cluster with 6 x 2 TB
hard drives?
Why is Cloudera Manager showing the concerning status KM the services?
You install Cloudera Manager on a cluster where each host has 1 GB of RAM. All of the services
show their status as concerning. However, all jobs submitted complete without an error.
Why is Cloudera Manager showing the concerning status KM the services?