Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
19 views4 pages

Internal Assesment Examination - Iv (Answer Key)

The document contains an internal assessment examination for a course on Grid and Cloud Computing. It includes questions on map-reduce paradigm, services offered in GT4, the purpose of heartbeats in Hadoop, and fault tolerance in HDFS. It also discusses HDFS architecture and data flow, MapReduce with examples, and provides details on Globus job workflow and the Globus Toolkit architecture. Finally, it gives notes on the Hadoop framework.

Uploaded by

Nithya Kandasamy
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
19 views4 pages

Internal Assesment Examination - Iv (Answer Key)

The document contains an internal assessment examination for a course on Grid and Cloud Computing. It includes questions on map-reduce paradigm, services offered in GT4, the purpose of heartbeats in Hadoop, and fault tolerance in HDFS. It also discusses HDFS architecture and data flow, MapReduce with examples, and provides details on Globus job workflow and the Globus Toolkit architecture. Finally, it gives notes on the Hadoop framework.

Uploaded by

Nithya Kandasamy
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

NANDHA COLLEGE OF TECHNOLOGY, ERODE - 52

DEPARTMENT OF CSE &IT


CS 6703- GRID AND CLOUD COMPUTING
INTERNAL ASSESMENT EXAMINATION – IV (Answer Key)
PART - A
1. How does divide-and-conquer strategy related to map-reduce paradigm?
The divide-and-conquer paradigm is often used to find an optimal solution of a problem. Its basic
idea is to decompose a given problem into two or more similar, but simpler, subproblems, to
solve them in turn, and to compose their solutions to solve the given problem.
2. Name any four services offered in GT4.
 security, resource management, job management, data services.
3. Write the significant use of GRAM.
The Gram stain is the most important staining procedure in microbiology. It is used to
differentiate between gram positive organisms and gram negative organisms. Hence, it is a
differential stain. Gram negative and gram positive organisms are distinguished from each other
by differences in their cell walls.
4. What is the purpose of heart beat in hadoop?
To make NameNode aware of the status(active / inactive) of DataNodes, each DataNode sends a
"Heart Beat Signal" for every 10 minutes(Default). This mechanism is called as HEART
BEAT MECHANISM. Based on this Heart Beat Signal Job Tracker assigns tasks to the Tasks
Trackers which are active.
5. “HDFS is fault tolerant”. Is it true? Justify your answer.
True. Fault tolerance in HDFS refers to the working strength of a system in unfavorable
conditions and how that system can handle such situation. HDFS is highly fault tolerant. It
handles faults by the process of replica creation. The replica of users data is created on different
machines in the HDFS cluster.
PART - B
6. a) Explain in detail about the Hadoop Distributed file system.(6)
 The Hadoop Distributed File System (HDFS) is the primary data storage system used
byHadoop applications.
 It employs a NameNode and DataNode architecture to implement adistributed file
system that provides high-performance access to data across highly
scalableHadoop clusters.

(7)

1
b) Discuss MAPREDUCE with suitable example. (4)
MapReduce is a programming framework that allows us to perform distributed and parallel
processing on large data sets in a distributed environment.
 Map (2)
 Redude (2)

(5)

7. a)Illustrate dataflow in HDFS during file read/write operation with suitable diagrams.(3)
 The Hadoop Core project provides the basic services for building a cloud computing
environment with commodity hardware, and the APIs for developing software that will run
on that cloud.
 The two fundamental pieces of Hadoop Core are the MapReduce framework, the cloud
computing environment, and he Hadoop Distributed File System (HDFS).
 The user is required to tell the framework the following:
• The location(s) in the distributed file system of the job input
• The location(s) in the distributed file system for the job output
• The input format
• The output format
• The class containing the map function
• Optionally. the class containing the reduce function

(5)

2
(5)

b) What is GT4? Describe in detail the Globus job work flow of GT4 with suitable
diagrams and illustration. (6)
 The Globus toolkit contains a set of libraries and programs that provides the developers
of specific tools or apps with solutions for common problems that are encountered when
creating a distributed system services and applications.
 A typical job execution sequence proceeds as follows: The user delegates his credentials
to a delegation service. The user submits a job request to GRAM with the delegation
identifier as a parameter. GRAM parses the request, retrieves the user proxy certificate
from the delegation service, and then acts on behalf of the user. GRAM sends a transfer
request to the RFT, which applies GridFTP to bring in the necessary files.

(7)

PART - C 1 x 14= 14 Marks

8. a) Explain in detail the global toolkit architecture with neat sketch.(7)


 The Globus Toolkit is an open middleware library for the grid computing communities.
These open source software libraries support many operational grids and their
applications on an international basis.

3
 The toolkit addresses common problems and issues related to grid resource discovery,
management, communication, security, fault detection, and portability. The software
itself provides a variety of components and capabilities.
 The library includes a rich set of service implementations. The implemented software
supports grid infrastructure management, provides tools for building new web services in
Java, C, and Python, builds a powerful standard-based.
 Security infrastructure and client APIs (in different languages), and offers comprehensive
command-line programs for accessing various grid services.
 The Globus Toolkit was initially motivated by a desire to remove obstacles that prevent
seamless collaboration, and thus sharing of resources and services, in scientific and
engineering applications.

(7)

b) Give a detailed note on hadoop framework.(7)


Hadoop is an Apache open source framework written in java that allows distributed processing of
large datasets across clusters of computers using simple programming models. The Hadoop
framework application works in an environment that provides
distributed storage and computation across clusters of computers.
 Hadoop Common − These are Java libraries and utilities required by other Hadoop modules.
 Hadoop YARN − This is a framework for job scheduling and cluster resource management.

(7)

You might also like