NPTEL Big Data Computing Week 2 Assignment Answers 2023

NPTEL Big Data Computing Week 2 Assignment Answers

1. What is the primary purpose of the Map phase in the MapReduce framework?

  • Combining and aggregating data.
  • Storing intermediate results.
  • Sorting and shuffling data.
  • Applying a user-defined function to each input record.
Answer :- For Answer Click Here

2. Which of the following statements about the components in the MapReduce framework is true?
Statement 1: The Job Tracker is hosted inside the master and it receives the job execution request from the client.
Statement 2: Task Tracker is the MapReduce component on the slave machine as there are multiple slave machines.

  • Both statements are true.
  • Only statement 1 is true.
  • Only statement 2 is true.
  • Both statements are false.
Answer :- For Answer Click Here

3. Which of the following is the slave/worker node and holds the user data in the form of Data Blocks?

  • NameNode
  • Data block
  • Replication
  • DataNode
Answer :- For Answer Click Here

4. The number of maps in MapReduce is usually driven by the total size of____________.

  • Inputs
  • Outputs
  • Tasks
  • None of the mentioned
Answer :- For Answer Click Here

5. Identify the correct statement(s) in the context of YARN (Yet Another Resource Negotiator):

A. YARN is highly scalable.
B. YARN enhances a Hadoop compute cluster in many ways.
C. YARN extends the power of Hadoop to incumbent and new technologies found within the data center.

Choose the correct option:
Only statement A is correct.
Statements A and B are correct.
Statements B and C are correct.
All statements A, B, and C are correct.

Answer :- For Answer Click Here

6. Which of the following statements accurately describe(s) the role and responsibilities of the Job Tracker in the context of Big Data computing?

A. The Job Tracker is hosted inside the master and it receives the job execution request from the client.
B. The Job Tracker breaks down big computations into smaller parts and allocates tasks to slave nodes.
C. The Job Tracker stores all the intermediate results from task execution on the master node.
D. The Job Tracker is responsible for managing the distributed file system in the cluster.

Choose the correct option:
Only statement A is correct.
Statements A and B are correct.
Statements A, B, and C are correct.
None of the statements are correct.

Answer :- For Answer Click Here

7. Consider the pseudo-code for MapReduce’s WordCount example. Let’s now assume that you want to determine the frequency of phrases consisting of 3 words each instead of determining the frequency of single words. Which part of the (pseudo-)code do you need to adapt?

  • Only map()
  • Only reduce()
  • map() and reduce()
  • None
Answer :- For Answer Click Here

8. How does the NameNode determine that a DataNode is active, using a mechanism known as:

  • Heartbeats
  • Datapulse
  • h-signal
  • Active-pulse
Answer :- For Answer Click Here

9. Which function processes a key/value pair to generate a set of intermediate key/value pairs?

  • Map
  • Reduce
  • Both Map and Reduce
  • None of the mentioned
Answer :- For Answer Click Here

10. Which of the following options correctly identifies the three main components of the YARN Scheduler in Hadoop?

  • Global Application Manager (GAM), Cluster Resource Tracker (CRT), Job Task Coordinator (JTC)
  • Resource Monitor (RM), Cluster Supervisor (CS), Task Executor (TE)
  • Global Resource Manager (RM), Per server Node Manager (NM), Per application (job) Application Master (AM)
  • Central Resource Coordinator (CRC), Node Resource Manager (NRM), Application Controller (AC)
Answer :- For Answer Click Here
Course NameBig Data Computing
CategoryNPTEL Assignment Answer
Home Click Here
Join Us on TelegramClick Here