
22AI62-BDA & CC MODULE 4 QUIZ
Quiz by Ramya K
Feel free to use or edit a copy
includes Teacher and Student dashboards
Measure skillsfrom any curriculum
Tag the questions with any skills you have. Your dashboard will track each student's mastery of each skill.
- edit the questions
- save a copy for later
- start a class game
- automatically assign follow-up activities based on students’ scores
- assign as homework
- share a link with colleagues
- print as a bubble sheet
- Q1
What does the Hadoop 'wordcount' example demonstrate?
File compression
Streaming input processing
Tokenizing and counting words
Sorting algorithms
30s - Q2
Which command runs a MapReduce example in Hadoop?
hadoop version
hadoop jar hadoop-examples.jar wordcount input output
hadoop fs -put
hadoop config
30s - Q3
The purpose of Hadoop benchmarks is to measure:
Data integrity
Cluster availability
Performance metrics
Security features
30s - Q4
Which phase sorts and groups intermediate data by key?
Shuffle
Combine
Map
Output
30s - Q5
MapReduce processes data in the form of:
Objects
XML files
Key-value pairs
Tuples
30s - Q6
Which component ensures job completion if a node fails?
JobTracker
HDFS
TaskTracker
Speculative execution
30s - Q7
What does speculative execution in Hadoop do?
Runs redundant copies of slow tasks
Combines reducers
Encrypts sensitive data
Runs jobs in parallel clusters
30s - Q8
Intermediate outputs in MapReduce are written to:
Cloud storage
Zookeeper
Local disk
HDFS
30s - Q9
In a WordCount program, the mapper emits:
Sentence and 1
Word and sentence
Word and 0
Word and 1
30s - Q10
Which language interface is used in Hadoop Streaming?
Python or Bash
Only Java
C++ only
C only
30s - Q11
Which of the following is required when using the Pipes interface?
C++ compiler
JVM
Python interpreter
Bash shell
30s - Q12
What is the role of the Context object in MapReduce?
Writes key-value pairs in map/reduce
Reads input files
Compiles source code
Tracks job progress
30s - Q13
A successful MapReduce job stores final output in:
STDOUT
Temporary directory
Local cache
Output directory on HDFS
30s - Q14
What happens if the output directory already exists in HDFS before running a job?
Job continues
Error is thrown
Files are overwritten
Directory is deleted
30s - Q15
Which file defines Hadoop job configuration?
core-site.xml
mapred-site.xml
job.xml
hdfs-site.xml
30s