Sail E0 Webinar

MCQs

Total Questions : 10
Question 1. Point out the wrong statement :
  1.    The Mapper outputs are sorted and then partitioned per Reducer
  2.    The total number of partitions is the same as the number of reduce tasks for the job
  3.    The intermediate, sorted outputs are always stored in a simple (key-len, key, value-len, value) format
  4.    None of the mentioned
 Discuss Question
Answer: Option D. -> None of the mentioned


All intermediate values associated with a given output key are subsequently grouped by the framework, and passed to the Reducer(s) to determine the final output.


Question 2. Point out the wrong statement :
  1.    It is legal to set the number of reduce-tasks to zero if no reduction is desired
  2.    The outputs of the map-tasks go directly to the FileSystem
  3.    The Mapreduce framework does not sort the map-outputs before writing them out to the FileSystem
  4.    None of the mentioned
 Discuss Question
Answer: Option D. -> None of the mentioned


Outputs of the map-tasks go directly to the FileSystem, into the output path set by setOutputPath(Path).


Question 3. Applications can use the ____________ to report progress and set application-level status messages
  1.    Partitioner
  2.    OutputSplit
  3.    Reporter
  4.    All of the mentioned
 Discuss Question
Answer: Option C. -> Reporter


Reporter is also used to update Counters, or just indicate that they are alive.


Question 4. __________ is the primary interface for a user to describe a MapReduce job to the Hadoop framework for execution.
  1.    JobConfig
  2.    JobConf
  3.    JobConfiguration
  4.    All of the mentioned
 Discuss Question
Answer: Option B. -> JobConf


JobConf is typically used to specify the Mapper, combiner (if any), Partitioner, Reducer, InputFormat, OutputFormat and OutputCommitter implementations.


Question 5. The right level of parallelism for maps seems to be around _________ maps per-node
  1.    1-10
  2.    10-100
  3.    100-150
  4.    150-200
 Discuss Question
Answer: Option B. -> 10-100


Task setup takes a while, so it is best if the maps take at least a minute to execute


Question 6. The ___________ executes the Mapper/ Reducer task as a child process in a separate jvm.
  1.    JobTracker
  2.    TaskTracker
  3.    TaskScheduler
  4.    None of the mentioned
 Discuss Question
Answer: Option A. -> JobTracker


The child-task inherits the environment of the parent TaskTracker.


Question 7. During the execution of a streaming job, the names of the _______ parameters are transformed.
  1.    vmap
  2.    mapvim
  3.    mapreduce
  4.    mapred
 Discuss Question
Answer: Option D. -> mapred


To get the values in a streaming job's mapper/reducer use the parameter names with the underscores.


Question 8. The ________ class provides the getValue() method to read the values from its instance.
  1.    Get
  2.    Result
  3.    Put
  4.    Value
 Discuss Question
Answer: Option B. -> Result


Get the result by passing your Get class instance to the get method of the HTable class. This method returns the Result class object, which holds the requested result.


Question 9. The standard output (stdout) and error (stderr) streams of the task are read by the TaskTracker and logged to :
  1.    ${HADOOP_LOG_DIR}/user
  2.    ${HADOOP_LOG_DIR}/userlogs
  3.    ${HADOOP_LOG_DIR}/logs
  4.    None of the mentioned
 Discuss Question
Answer: Option B. -> ${HADOOP_LOG_DIR}/userlogs


The child-jvm always has its current working directory added to the java.library.path and LD_LIBRARY_PATH.


Question 10. __________ class adds HBase configuration files to its object.
  1.    Configuration
  2.    Collector
  3.    Component
  4.    None of the mentioned
 Discuss Question
Answer: Option A. -> Configuration


You can create a configuration object using the create() method of the HbaseConfiguration class.


Latest Videos

Latest Test Papers