NPTEL Big Data Computing Week 1 Assignment Answers 2023

Join Our WhatsApp Group Join Now
Join Us On Telegram Join Now

NPTEL Big Data Computing Week 1 Assignment Solution

NPTEL Big Data Computing Week 1 Assignment Answers 2023

NPTEL Big Data Computing Week 1 Assignment Answers

1. What are the three key characteristics of Big Data, often referred to as the 3V’s, according to IBM?

  • Viscosity, Velocity, Veracity
  • Volume, Value, Variety
  • Volume, Velocity, Variety
  • Volumetric, Visceral, Vortex
Answer :- For Answer Click Here

2. What is the primary purpose of the MapReduce programming model in processing and generating large data sets?

  • To directly process and analyze data without any intermediate steps.
  • To convert unstructured data into structured data.
  • To specify a map function for generating intermediate key/value pairs and a reduce function for merging values associated with the same key.
  • To create visualizations and graphs for large data sets.
Answer :- For Answer Click Here

3. _____ is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data.

  • Flume
  • Apache Sqoop
  • Pig
  • Mahout
Answer :- For Answer Click Here

4. What is the primary role of YARN (Yet Another Resource Manager) in the Apache Hadoop ecosystem?

  • YARN is a data storage layer for managing and storing large datasets in Hadoop clusters.
  • YARN is a programming model for processing and analyzing data in Hadoop clusters.
  • YARN is responsible for allocating system resources and scheduling tasks for applications in a Hadoop cluster.
  • YARN is a visualization tool for creating graphs and charts based on Hadoop data.
Answer :- For Answer Click Here

5. Which of the following statements accurately describes the characteristics and functionality of HDFS (Hadoop Distributed File System)?

  • HDFS is a centralized file system designed for storing small files and achieving high-speed data processing.
  • HDFS is a programming language used for writing MapReduce applications within the Hadoop ecosystem.
  • HDFS is a distributed, scalable, and portable file system designed for storing large files across multiple machines, achieving reliability through replication.
  • HDFS is a visualization tool that generates graphs and charts based on data stored in the Hadoop ecosystem.
Answer :- For Answer Click Here

6. Which statement accurately describes the role and design of HBase in the Hadoop stack?

  • HBase is a programming language used for writing complex data processing algorithms in the Hadoop ecosystem.
  • HBase is a data warehousing solution designed for batch processing of large datasets in Hadoop clusters.
  • HBase is a key-value store that provides fast random access to substantial datasets, making it suitable for applications requiring such access patterns.
  • HBase is a visualization tool that generates charts and graphs based on data stored in Hadoop clusters.
Answer :- For Answer Click Here

7. ______ brings scalable parallel database technology to Hadoop and allows users to submit low latencies queries to the data that’s stored within the HDFS or the Hbase without acquiring a ton of data movement and manipulation.

  • Apache Sqoop
  • Mahout
  • Flume
  • Impala
Answer :- For Answer Click Here

8. What is the primary purpose of ZooKeeper in a distributed system?

  • ZooKeeper is a data warehousing solution for storing and managing large datasets in a distributed cluster.
  • ZooKeeper is a programming language for developing distributed applications in a cloud environment.
  • ZooKeeper is a highly reliable distributed coordination kernel used for tasks such as distributed locking, configuration management, leadership election, and work queues.
  • ZooKeeper is a visualization tool for creating graphs and charts based on data stored in distributed systems.
Answer :- For Answer Click Here

9. ____ is a distributed file system that stores data on a commodity machine. Providing very high aggregate bandwidth across the entire cluster.

  • Hadoop Common
  • Hadoop Distributed File System (HDFS)
  • Hadoop YARN
  • Hadoop MapReduce
Answer :- For Answer Click Here

10. Which statement accurately describes Spark MLlib?

  • Spark MLlib is a visualization tool for creating charts and graphs based on data processed in Spark clusters.
  • Spark MLlib is a programming language used for writing Spark applications in a distributed environment.
  • Spark MLlib is a distributed machine learning framework built on top of Spark Core, providing scalable machine learning algorithms and utilities for tasks such as classification, regression, clustering, and collaborative filtering.
  • Spark MLlib is a data warehousing solution for storing and querying large datasets in a Spark cluster.
Answer :- For Answer Click Here
Course NameBig Data Computing
CategoryNPTEL Assignment Answer
Home Click Here
Join Us on TelegramClick Here

Leave a comment